var/home/core/zuul-output/0000755000175000017500000000000015112433242014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015112445340015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005242713315112445332017704 0ustar rootrootNov 28 23:59:45 crc systemd[1]: Starting Kubernetes Kubelet... Nov 28 23:59:45 crc restorecon[4690]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:45 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 23:59:46 crc restorecon[4690]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 28 23:59:46 crc restorecon[4690]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 28 23:59:46 crc kubenswrapper[4931]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 23:59:46 crc kubenswrapper[4931]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 28 23:59:46 crc kubenswrapper[4931]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 23:59:46 crc kubenswrapper[4931]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 23:59:46 crc kubenswrapper[4931]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 28 23:59:46 crc kubenswrapper[4931]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.963623 4931 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966389 4931 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966408 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966413 4931 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966417 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966422 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966426 4931 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966430 4931 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966434 4931 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966443 4931 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966449 4931 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966454 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966459 4931 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966464 4931 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966469 4931 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966475 4931 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966480 4931 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966483 4931 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966488 4931 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966494 4931 feature_gate.go:330] unrecognized feature gate: Example Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966499 4931 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966503 4931 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966508 4931 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966512 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966516 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966521 4931 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966526 4931 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966531 4931 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966536 4931 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966540 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966544 4931 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966549 4931 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966554 4931 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966559 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966563 4931 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966568 4931 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966572 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966577 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966583 4931 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966588 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966593 4931 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966598 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966602 4931 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966607 4931 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966611 4931 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966616 4931 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966620 4931 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966627 4931 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966634 4931 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966638 4931 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966642 4931 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966645 4931 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966649 4931 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966653 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966656 4931 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966661 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966664 4931 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966668 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966671 4931 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966675 4931 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966680 4931 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966684 4931 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966687 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966691 4931 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966695 4931 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966699 4931 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966703 4931 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966707 4931 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966711 4931 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966715 4931 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966718 4931 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.966723 4931 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967329 4931 flags.go:64] FLAG: --address="0.0.0.0" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967345 4931 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967356 4931 flags.go:64] FLAG: --anonymous-auth="true" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967363 4931 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967371 4931 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967376 4931 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967382 4931 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967388 4931 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967393 4931 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967399 4931 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967404 4931 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967408 4931 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967412 4931 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967417 4931 flags.go:64] FLAG: --cgroup-root="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967421 4931 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967425 4931 flags.go:64] FLAG: --client-ca-file="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967429 4931 flags.go:64] FLAG: --cloud-config="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967433 4931 flags.go:64] FLAG: --cloud-provider="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967438 4931 flags.go:64] FLAG: --cluster-dns="[]" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967443 4931 flags.go:64] FLAG: --cluster-domain="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967447 4931 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967452 4931 flags.go:64] FLAG: --config-dir="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967456 4931 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967461 4931 flags.go:64] FLAG: --container-log-max-files="5" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967468 4931 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967472 4931 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967477 4931 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967481 4931 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967486 4931 flags.go:64] FLAG: --contention-profiling="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967491 4931 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967496 4931 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967501 4931 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967505 4931 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967511 4931 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967515 4931 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967519 4931 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967524 4931 flags.go:64] FLAG: --enable-load-reader="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967528 4931 flags.go:64] FLAG: --enable-server="true" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967532 4931 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967539 4931 flags.go:64] FLAG: --event-burst="100" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967544 4931 flags.go:64] FLAG: --event-qps="50" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967548 4931 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967552 4931 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967557 4931 flags.go:64] FLAG: --eviction-hard="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967562 4931 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967567 4931 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967572 4931 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967576 4931 flags.go:64] FLAG: --eviction-soft="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967580 4931 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967584 4931 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967589 4931 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967594 4931 flags.go:64] FLAG: --experimental-mounter-path="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967598 4931 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967602 4931 flags.go:64] FLAG: --fail-swap-on="true" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967606 4931 flags.go:64] FLAG: --feature-gates="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967612 4931 flags.go:64] FLAG: --file-check-frequency="20s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967617 4931 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967621 4931 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967626 4931 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967630 4931 flags.go:64] FLAG: --healthz-port="10248" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967635 4931 flags.go:64] FLAG: --help="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967639 4931 flags.go:64] FLAG: --hostname-override="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967648 4931 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967653 4931 flags.go:64] FLAG: --http-check-frequency="20s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967658 4931 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967662 4931 flags.go:64] FLAG: --image-credential-provider-config="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967666 4931 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967670 4931 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967675 4931 flags.go:64] FLAG: --image-service-endpoint="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967679 4931 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967683 4931 flags.go:64] FLAG: --kube-api-burst="100" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967687 4931 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967692 4931 flags.go:64] FLAG: --kube-api-qps="50" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967696 4931 flags.go:64] FLAG: --kube-reserved="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967700 4931 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967704 4931 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967709 4931 flags.go:64] FLAG: --kubelet-cgroups="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967713 4931 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967717 4931 flags.go:64] FLAG: --lock-file="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967721 4931 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967725 4931 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967730 4931 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967737 4931 flags.go:64] FLAG: --log-json-split-stream="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967741 4931 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967746 4931 flags.go:64] FLAG: --log-text-split-stream="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967750 4931 flags.go:64] FLAG: --logging-format="text" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967754 4931 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967759 4931 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967764 4931 flags.go:64] FLAG: --manifest-url="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967769 4931 flags.go:64] FLAG: --manifest-url-header="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967776 4931 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967781 4931 flags.go:64] FLAG: --max-open-files="1000000" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967788 4931 flags.go:64] FLAG: --max-pods="110" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967792 4931 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967799 4931 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967826 4931 flags.go:64] FLAG: --memory-manager-policy="None" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967831 4931 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967837 4931 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967842 4931 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967847 4931 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967859 4931 flags.go:64] FLAG: --node-status-max-images="50" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967864 4931 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967869 4931 flags.go:64] FLAG: --oom-score-adj="-999" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967874 4931 flags.go:64] FLAG: --pod-cidr="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967879 4931 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967887 4931 flags.go:64] FLAG: --pod-manifest-path="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967892 4931 flags.go:64] FLAG: --pod-max-pids="-1" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967897 4931 flags.go:64] FLAG: --pods-per-core="0" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967902 4931 flags.go:64] FLAG: --port="10250" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967906 4931 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967911 4931 flags.go:64] FLAG: --provider-id="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967916 4931 flags.go:64] FLAG: --qos-reserved="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967921 4931 flags.go:64] FLAG: --read-only-port="10255" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967926 4931 flags.go:64] FLAG: --register-node="true" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967931 4931 flags.go:64] FLAG: --register-schedulable="true" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967936 4931 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967943 4931 flags.go:64] FLAG: --registry-burst="10" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967949 4931 flags.go:64] FLAG: --registry-qps="5" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967954 4931 flags.go:64] FLAG: --reserved-cpus="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967958 4931 flags.go:64] FLAG: --reserved-memory="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967965 4931 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967975 4931 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967981 4931 flags.go:64] FLAG: --rotate-certificates="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967985 4931 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967990 4931 flags.go:64] FLAG: --runonce="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.967995 4931 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968002 4931 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968008 4931 flags.go:64] FLAG: --seccomp-default="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968012 4931 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968017 4931 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968022 4931 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968026 4931 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968031 4931 flags.go:64] FLAG: --storage-driver-password="root" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968035 4931 flags.go:64] FLAG: --storage-driver-secure="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968039 4931 flags.go:64] FLAG: --storage-driver-table="stats" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968044 4931 flags.go:64] FLAG: --storage-driver-user="root" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968048 4931 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968053 4931 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968057 4931 flags.go:64] FLAG: --system-cgroups="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968061 4931 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968068 4931 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968072 4931 flags.go:64] FLAG: --tls-cert-file="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968076 4931 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968082 4931 flags.go:64] FLAG: --tls-min-version="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968086 4931 flags.go:64] FLAG: --tls-private-key-file="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968090 4931 flags.go:64] FLAG: --topology-manager-policy="none" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968094 4931 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968098 4931 flags.go:64] FLAG: --topology-manager-scope="container" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968103 4931 flags.go:64] FLAG: --v="2" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968110 4931 flags.go:64] FLAG: --version="false" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968117 4931 flags.go:64] FLAG: --vmodule="" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968124 4931 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968129 4931 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968247 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968253 4931 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968258 4931 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968263 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968268 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968275 4931 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968279 4931 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968284 4931 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968289 4931 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968293 4931 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968297 4931 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968301 4931 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968305 4931 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968309 4931 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968313 4931 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968317 4931 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968321 4931 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968324 4931 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968328 4931 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968332 4931 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968336 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968340 4931 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968344 4931 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968348 4931 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968351 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968355 4931 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968359 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968363 4931 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968366 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968370 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968374 4931 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968378 4931 feature_gate.go:330] unrecognized feature gate: Example Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968384 4931 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968388 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968392 4931 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968395 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968400 4931 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968407 4931 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968410 4931 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968414 4931 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968418 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968422 4931 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968425 4931 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968429 4931 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968433 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968437 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968441 4931 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968445 4931 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968449 4931 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968452 4931 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968456 4931 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968460 4931 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968464 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968467 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968471 4931 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968475 4931 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968479 4931 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968482 4931 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968486 4931 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968490 4931 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968494 4931 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968499 4931 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968504 4931 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968508 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968514 4931 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968519 4931 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968523 4931 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968527 4931 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968532 4931 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968540 4931 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.968544 4931 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.968701 4931 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.983954 4931 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.984572 4931 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984767 4931 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984791 4931 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984802 4931 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984845 4931 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984856 4931 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984865 4931 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984879 4931 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984891 4931 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984901 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984910 4931 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984918 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984926 4931 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984934 4931 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984943 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984952 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984960 4931 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984968 4931 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984976 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984984 4931 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.984996 4931 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985004 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985012 4931 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985021 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985029 4931 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985039 4931 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985048 4931 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985057 4931 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985069 4931 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985084 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985093 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985105 4931 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985114 4931 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985126 4931 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985136 4931 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985145 4931 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985154 4931 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985163 4931 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985171 4931 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985182 4931 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985192 4931 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985201 4931 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985210 4931 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985219 4931 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985227 4931 feature_gate.go:330] unrecognized feature gate: Example Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985236 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985244 4931 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985252 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985260 4931 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985269 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985277 4931 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985285 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985292 4931 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985300 4931 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985308 4931 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985316 4931 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985326 4931 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985337 4931 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985347 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985356 4931 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985364 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985373 4931 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985381 4931 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985389 4931 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985396 4931 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985404 4931 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985412 4931 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985420 4931 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985428 4931 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985435 4931 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985443 4931 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985451 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.985465 4931 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985735 4931 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985751 4931 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985760 4931 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985769 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985778 4931 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985785 4931 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985793 4931 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985801 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985833 4931 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985842 4931 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985850 4931 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985861 4931 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985873 4931 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985883 4931 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985893 4931 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985903 4931 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985913 4931 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985921 4931 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985929 4931 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985937 4931 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985945 4931 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985953 4931 feature_gate.go:330] unrecognized feature gate: Example Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985960 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985968 4931 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985976 4931 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985985 4931 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.985992 4931 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986001 4931 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986011 4931 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986021 4931 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986030 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986040 4931 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986049 4931 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986059 4931 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986068 4931 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986078 4931 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986087 4931 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986099 4931 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986108 4931 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986116 4931 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986124 4931 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986133 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986141 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986149 4931 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986157 4931 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986164 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986173 4931 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986181 4931 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986190 4931 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986198 4931 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986205 4931 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986213 4931 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986221 4931 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986228 4931 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986237 4931 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986244 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986252 4931 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986260 4931 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986267 4931 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986275 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986283 4931 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986291 4931 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986298 4931 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986306 4931 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986316 4931 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986325 4931 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986333 4931 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986341 4931 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986349 4931 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986356 4931 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 28 23:59:46 crc kubenswrapper[4931]: W1128 23:59:46.986364 4931 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.986376 4931 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.986696 4931 server.go:940] "Client rotation is on, will bootstrap in background" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.991600 4931 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.991777 4931 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.992729 4931 server.go:997] "Starting client certificate rotation" Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.992784 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.993334 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-12 14:23:33.842927218 +0000 UTC Nov 28 23:59:46 crc kubenswrapper[4931]: I1128 23:59:46.993499 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.000602 4931 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.004023 4931 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.004956 4931 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.016702 4931 log.go:25] "Validated CRI v1 runtime API" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.078105 4931 log.go:25] "Validated CRI v1 image API" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.080535 4931 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.083687 4931 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-28-23-55-24-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.083747 4931 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.113885 4931 manager.go:217] Machine: {Timestamp:2025-11-28 23:59:47.111695423 +0000 UTC m=+0.273588745 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:dfbef0e2-70ff-4485-b415-c8232fbdbdee BootID:fc84002b-470f-4b68-bb6c-c1b6bbb7e873 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:8c:7d:9f Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:8c:7d:9f Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:34:dd:11 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:46:1f:00 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:d1:53:3c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:50:91:f0 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:26:fb:0e:0b:10:65 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:f2:40:6e:0b:87:fb Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.114293 4931 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.114614 4931 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.115310 4931 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.115691 4931 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.115766 4931 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.116277 4931 topology_manager.go:138] "Creating topology manager with none policy" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.116304 4931 container_manager_linux.go:303] "Creating device plugin manager" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.116665 4931 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.116748 4931 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.117385 4931 state_mem.go:36] "Initialized new in-memory state store" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.117568 4931 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.118624 4931 kubelet.go:418] "Attempting to sync node with API server" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.118665 4931 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.118722 4931 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.118752 4931 kubelet.go:324] "Adding apiserver pod source" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.118778 4931 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 28 23:59:47 crc kubenswrapper[4931]: W1128 23:59:47.121487 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Nov 28 23:59:47 crc kubenswrapper[4931]: W1128 23:59:47.121493 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.121631 4931 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.121704 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.121595 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.122300 4931 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.123542 4931 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.124384 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.124431 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.124448 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.124466 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.124499 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.124517 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.124536 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.124560 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.124580 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.124596 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.124644 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.124660 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.125176 4931 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.126105 4931 server.go:1280] "Started kubelet" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.126495 4931 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.126609 4931 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.127253 4931 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.127510 4931 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 28 23:59:47 crc systemd[1]: Started Kubernetes Kubelet. Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.130034 4931 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.130378 4931 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.130490 4931 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 02:27:38.588198341 +0000 UTC Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.131661 4931 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.131713 4931 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.131796 4931 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.131543 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="200ms" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.130939 4931 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.132462 4931 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.18:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187c512833719832 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-28 23:59:47.12601605 +0000 UTC m=+0.287909342,LastTimestamp:2025-11-28 23:59:47.12601605 +0000 UTC m=+0.287909342,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 28 23:59:47 crc kubenswrapper[4931]: W1128 23:59:47.133343 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.133458 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.133511 4931 factory.go:55] Registering systemd factory Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.133551 4931 factory.go:221] Registration of the systemd container factory successfully Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.135358 4931 factory.go:153] Registering CRI-O factory Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.135895 4931 factory.go:221] Registration of the crio container factory successfully Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.136029 4931 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.136073 4931 factory.go:103] Registering Raw factory Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.136107 4931 manager.go:1196] Started watching for new ooms in manager Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.137330 4931 manager.go:319] Starting recovery of all containers Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.141551 4931 server.go:460] "Adding debug handlers to kubelet server" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.149733 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.149884 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.149909 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.149930 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.149950 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.149970 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.149989 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150009 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150036 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150058 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150080 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150101 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150119 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150147 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150167 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150251 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150275 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150297 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150320 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150340 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.150361 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151301 4931 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151364 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151388 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151408 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151429 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151449 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151480 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151503 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151522 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151541 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151620 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151640 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151664 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151689 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151768 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151800 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151854 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151874 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151894 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151914 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151934 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151959 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.151979 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152007 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152027 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152045 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152066 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152085 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152104 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152124 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152145 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152164 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152195 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152216 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152237 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152262 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152284 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152305 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152327 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152349 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152371 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152392 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152415 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152437 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152459 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152480 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152502 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152522 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152544 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152566 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152588 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152609 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152634 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152655 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152679 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152701 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152724 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152743 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152779 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152804 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152859 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152882 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152904 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152925 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152948 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152968 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.152992 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153016 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153049 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153071 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153093 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153117 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153138 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153162 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153189 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153211 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153233 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153255 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153275 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153296 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153316 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153337 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153358 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153377 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153410 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153435 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153458 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153480 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153507 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153529 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153553 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153574 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153649 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153671 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153697 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153720 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153739 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153763 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153785 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153829 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153851 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153872 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153893 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153917 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153942 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.153971 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154006 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154030 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154052 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154074 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154097 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154120 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154141 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154164 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154187 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154209 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154231 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154251 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154272 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154292 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154314 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154335 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154359 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154381 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154401 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154426 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154446 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154468 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154489 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154512 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154534 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154556 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154582 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154603 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154625 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154651 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154671 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154690 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154713 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154734 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154757 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154782 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154834 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154864 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154887 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154907 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154928 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.154950 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155000 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155024 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155053 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155086 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155117 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155145 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155170 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155191 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155210 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155231 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155251 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155269 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155288 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155307 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155327 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155347 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155366 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155386 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155405 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155426 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155446 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155467 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155488 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155510 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155530 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155551 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155571 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155595 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155617 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155635 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155653 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155671 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155690 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155709 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155729 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155748 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155769 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155789 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155838 4931 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155855 4931 reconstruct.go:97] "Volume reconstruction finished" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.155870 4931 reconciler.go:26] "Reconciler: start to sync state" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.173152 4931 manager.go:324] Recovery completed Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.189547 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.192593 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.192672 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.192854 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.194236 4931 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.194273 4931 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.194321 4931 state_mem.go:36] "Initialized new in-memory state store" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.208194 4931 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.210874 4931 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.210952 4931 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.210998 4931 kubelet.go:2335] "Starting kubelet main sync loop" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.211069 4931 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 28 23:59:47 crc kubenswrapper[4931]: W1128 23:59:47.213032 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.213171 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.232376 4931 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.311333 4931 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.332537 4931 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.333240 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="400ms" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.433909 4931 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.459985 4931 policy_none.go:49] "None policy: Start" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.462469 4931 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.462521 4931 state_mem.go:35] "Initializing new in-memory state store" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.511646 4931 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.532436 4931 manager.go:334] "Starting Device Plugin manager" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.532524 4931 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.532548 4931 server.go:79] "Starting device plugin registration server" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.533388 4931 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.533432 4931 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.534193 4931 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.534373 4931 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.534408 4931 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.545213 4931 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.634515 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.636283 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.636338 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.636360 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.636421 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.637167 4931 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.18:6443: connect: connection refused" node="crc" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.735178 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="800ms" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.838080 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.839964 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.840031 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.840051 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.840100 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 23:59:47 crc kubenswrapper[4931]: E1128 23:59:47.840970 4931 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.18:6443: connect: connection refused" node="crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.912910 4931 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.913116 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.914931 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.914998 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.915021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.915275 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.915646 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.915712 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.916422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.916486 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.916513 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.916889 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.916936 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.916973 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.917331 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.917394 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.917419 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.918278 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.918349 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.918367 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.918710 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.918732 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.918747 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.918963 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.919241 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.919347 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.920096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.920159 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.920179 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.920391 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.920581 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.920639 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.921350 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.921396 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.921415 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.921928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.921973 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.921994 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.922027 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.922004 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.922127 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.922398 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.922473 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.924036 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.924279 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.924361 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966162 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966309 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966358 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966396 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966564 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966615 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966671 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966703 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966737 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966769 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966854 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966887 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966917 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966948 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:47 crc kubenswrapper[4931]: I1128 23:59:47.966978 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068278 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068356 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068406 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068472 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068518 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068565 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068612 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068659 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068706 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068752 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068797 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068896 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068888 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069055 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069102 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069151 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069191 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.068942 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069166 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069236 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069252 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069238 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069280 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069341 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069300 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069129 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069280 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069307 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069463 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.069604 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.127542 4931 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.130791 4931 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 08:19:33.491021031 +0000 UTC Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.130895 4931 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1064h19m45.360130306s for next certificate rotation Nov 28 23:59:48 crc kubenswrapper[4931]: W1128 23:59:48.224472 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Nov 28 23:59:48 crc kubenswrapper[4931]: E1128 23:59:48.224612 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.241442 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.242927 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.242987 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.243008 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.243053 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 23:59:48 crc kubenswrapper[4931]: E1128 23:59:48.244054 4931 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.18:6443: connect: connection refused" node="crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.266197 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.273584 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.299091 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: W1128 23:59:48.299766 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-182e8baa89875d7ab25c32e3e0600e3c74144cbb4d9de8107b8eafa4264c9756 WatchSource:0}: Error finding container 182e8baa89875d7ab25c32e3e0600e3c74144cbb4d9de8107b8eafa4264c9756: Status 404 returned error can't find the container with id 182e8baa89875d7ab25c32e3e0600e3c74144cbb4d9de8107b8eafa4264c9756 Nov 28 23:59:48 crc kubenswrapper[4931]: W1128 23:59:48.302430 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-bfdc39ce78a6e87929e8489afb0909612fa2c929ffd3a0a04a4c1274a6b28279 WatchSource:0}: Error finding container bfdc39ce78a6e87929e8489afb0909612fa2c929ffd3a0a04a4c1274a6b28279: Status 404 returned error can't find the container with id bfdc39ce78a6e87929e8489afb0909612fa2c929ffd3a0a04a4c1274a6b28279 Nov 28 23:59:48 crc kubenswrapper[4931]: W1128 23:59:48.315194 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-a714c42c301e9fb7a8be91181437ef1a746ef894942831510d6180f492a964de WatchSource:0}: Error finding container a714c42c301e9fb7a8be91181437ef1a746ef894942831510d6180f492a964de: Status 404 returned error can't find the container with id a714c42c301e9fb7a8be91181437ef1a746ef894942831510d6180f492a964de Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.322985 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: I1128 23:59:48.332897 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:48 crc kubenswrapper[4931]: W1128 23:59:48.353905 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-1f552214760a0af06133dd3aa8e8dbd1b29daba5f7be9291f3877d4e4c1a04e8 WatchSource:0}: Error finding container 1f552214760a0af06133dd3aa8e8dbd1b29daba5f7be9291f3877d4e4c1a04e8: Status 404 returned error can't find the container with id 1f552214760a0af06133dd3aa8e8dbd1b29daba5f7be9291f3877d4e4c1a04e8 Nov 28 23:59:48 crc kubenswrapper[4931]: E1128 23:59:48.536082 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="1.6s" Nov 28 23:59:48 crc kubenswrapper[4931]: W1128 23:59:48.618133 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Nov 28 23:59:48 crc kubenswrapper[4931]: E1128 23:59:48.618262 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Nov 28 23:59:48 crc kubenswrapper[4931]: W1128 23:59:48.695028 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Nov 28 23:59:48 crc kubenswrapper[4931]: E1128 23:59:48.695153 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Nov 28 23:59:48 crc kubenswrapper[4931]: W1128 23:59:48.711197 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Nov 28 23:59:48 crc kubenswrapper[4931]: E1128 23:59:48.711276 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.044334 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.047732 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.047841 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.047864 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.047913 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 23:59:49 crc kubenswrapper[4931]: E1128 23:59:49.048612 4931 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.18:6443: connect: connection refused" node="crc" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.104423 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 28 23:59:49 crc kubenswrapper[4931]: E1128 23:59:49.106211 4931 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.127770 4931 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.225101 4931 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264" exitCode=0 Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.225214 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264"} Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.225346 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"bfdc39ce78a6e87929e8489afb0909612fa2c929ffd3a0a04a4c1274a6b28279"} Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.225454 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.226581 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.226621 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.226633 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.229659 4931 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52" exitCode=0 Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.229732 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52"} Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.229864 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"182e8baa89875d7ab25c32e3e0600e3c74144cbb4d9de8107b8eafa4264c9756"} Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.230028 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.231865 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.231913 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.231933 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.233207 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c"} Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.233255 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1f552214760a0af06133dd3aa8e8dbd1b29daba5f7be9291f3877d4e4c1a04e8"} Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.236582 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21" exitCode=0 Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.236667 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21"} Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.236697 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9b923829b5ea24b8a278db930e39bf288c09e55931ff9e5a9776b336616b36ca"} Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.236889 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.237945 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.237988 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.238006 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.240288 4931 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57" exitCode=0 Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.240332 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57"} Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.240359 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a714c42c301e9fb7a8be91181437ef1a746ef894942831510d6180f492a964de"} Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.240495 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.241575 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.241601 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.241621 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.242573 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.243613 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.243681 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:49 crc kubenswrapper[4931]: I1128 23:59:49.243712 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.249101 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e53bafde2aa1cbcfcdb36a53eabbf234b09505c13a764ae7ef41659856767dfc"} Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.249258 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.251910 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.251953 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.252033 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.255432 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6b89687792ce16c3406e8500eb2bfa867fcdc1db444a0623caca0b5033229f33"} Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.255462 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.255483 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a9dc7cc6f5295935a3444f30c164cc63f6b2639a70fd4f45098f28ae9d0f98d2"} Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.255497 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9c38778155a771c99ed4c71c12be76871dc3299129f77a1551fc19f8482e4065"} Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.256405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.256436 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.256447 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.261957 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a"} Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.262031 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e"} Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.262045 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794"} Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.262136 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.263566 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.263612 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.263626 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.266892 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8"} Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.266924 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c"} Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.266938 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082"} Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.266950 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975"} Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.268802 4931 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087" exitCode=0 Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.268850 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087"} Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.268969 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.269911 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.269937 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.269945 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.649727 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.651599 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.651674 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.651685 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:50 crc kubenswrapper[4931]: I1128 23:59:50.651719 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.277766 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e"} Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.277975 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.280186 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.280260 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.280281 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.283034 4931 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd" exitCode=0 Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.283125 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd"} Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.283264 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.283321 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.283354 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.283362 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.288763 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.288909 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.288948 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.289054 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.289114 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.289164 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.289657 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.289717 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.289757 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.657198 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.663538 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:51 crc kubenswrapper[4931]: I1128 23:59:51.857621 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.292459 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219"} Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.292533 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470"} Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.292556 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209"} Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.292556 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.292659 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.292759 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.294107 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.294172 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.294193 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.294583 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.294640 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.294656 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:52 crc kubenswrapper[4931]: I1128 23:59:52.790584 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:52.999930 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.000315 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.002203 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.002281 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.002304 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.308476 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.308581 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.309109 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967"} Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.309177 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a"} Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.309387 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.310432 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.310487 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.310508 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.311116 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.311205 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.311240 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.386197 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.460555 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.460779 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.460872 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.462539 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.462608 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:53 crc kubenswrapper[4931]: I1128 23:59:53.462623 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.316903 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.317111 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.317300 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.319701 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.319772 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.319793 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.319922 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.319979 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.320001 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.705879 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.706416 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.706861 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.708860 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.708981 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.709010 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.858049 4931 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 28 23:59:54 crc kubenswrapper[4931]: I1128 23:59:54.858196 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 28 23:59:55 crc kubenswrapper[4931]: I1128 23:59:55.195300 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 28 23:59:55 crc kubenswrapper[4931]: I1128 23:59:55.320487 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:55 crc kubenswrapper[4931]: I1128 23:59:55.322160 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:55 crc kubenswrapper[4931]: I1128 23:59:55.322211 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:55 crc kubenswrapper[4931]: I1128 23:59:55.322236 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:56 crc kubenswrapper[4931]: I1128 23:59:56.161700 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 28 23:59:56 crc kubenswrapper[4931]: I1128 23:59:56.162525 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:56 crc kubenswrapper[4931]: I1128 23:59:56.164609 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:56 crc kubenswrapper[4931]: I1128 23:59:56.164746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:56 crc kubenswrapper[4931]: I1128 23:59:56.164778 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:57 crc kubenswrapper[4931]: I1128 23:59:57.205607 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 28 23:59:57 crc kubenswrapper[4931]: I1128 23:59:57.205894 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:57 crc kubenswrapper[4931]: I1128 23:59:57.207720 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:57 crc kubenswrapper[4931]: I1128 23:59:57.207770 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:57 crc kubenswrapper[4931]: I1128 23:59:57.207784 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 28 23:59:57 crc kubenswrapper[4931]: E1128 23:59:57.545466 4931 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 28 23:59:58 crc kubenswrapper[4931]: I1128 23:59:58.966208 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 28 23:59:58 crc kubenswrapper[4931]: I1128 23:59:58.966492 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 28 23:59:58 crc kubenswrapper[4931]: I1128 23:59:58.968122 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 28 23:59:58 crc kubenswrapper[4931]: I1128 23:59:58.968302 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 28 23:59:58 crc kubenswrapper[4931]: I1128 23:59:58.968442 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:00 crc kubenswrapper[4931]: W1129 00:00:00.097205 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 29 00:00:00 crc kubenswrapper[4931]: I1129 00:00:00.097320 4931 trace.go:236] Trace[181512188]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 23:59:50.096) (total time: 10001ms): Nov 29 00:00:00 crc kubenswrapper[4931]: Trace[181512188]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (00:00:00.097) Nov 29 00:00:00 crc kubenswrapper[4931]: Trace[181512188]: [10.001135822s] [10.001135822s] END Nov 29 00:00:00 crc kubenswrapper[4931]: E1129 00:00:00.097347 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 29 00:00:00 crc kubenswrapper[4931]: I1129 00:00:00.129094 4931 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 29 00:00:00 crc kubenswrapper[4931]: E1129 00:00:00.137469 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Nov 29 00:00:00 crc kubenswrapper[4931]: W1129 00:00:00.618787 4931 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 29 00:00:00 crc kubenswrapper[4931]: I1129 00:00:00.619041 4931 trace.go:236] Trace[1635362179]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 23:59:50.617) (total time: 10001ms): Nov 29 00:00:00 crc kubenswrapper[4931]: Trace[1635362179]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (00:00:00.618) Nov 29 00:00:00 crc kubenswrapper[4931]: Trace[1635362179]: [10.001857987s] [10.001857987s] END Nov 29 00:00:00 crc kubenswrapper[4931]: E1129 00:00:00.619079 4931 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 29 00:00:00 crc kubenswrapper[4931]: E1129 00:00:00.653084 4931 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Nov 29 00:00:00 crc kubenswrapper[4931]: I1129 00:00:00.950752 4931 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 29 00:00:00 crc kubenswrapper[4931]: I1129 00:00:00.950853 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 29 00:00:00 crc kubenswrapper[4931]: I1129 00:00:00.965138 4931 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 29 00:00:00 crc kubenswrapper[4931]: I1129 00:00:00.965277 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 29 00:00:03 crc kubenswrapper[4931]: I1129 00:00:03.853648 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 00:00:03 crc kubenswrapper[4931]: I1129 00:00:03.855093 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:03 crc kubenswrapper[4931]: I1129 00:00:03.855132 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:03 crc kubenswrapper[4931]: I1129 00:00:03.855155 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:03 crc kubenswrapper[4931]: I1129 00:00:03.855182 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 29 00:00:03 crc kubenswrapper[4931]: E1129 00:00:03.860585 4931 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 29 00:00:04 crc kubenswrapper[4931]: I1129 00:00:04.027037 4931 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 29 00:00:04 crc kubenswrapper[4931]: I1129 00:00:04.714740 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:00:04 crc kubenswrapper[4931]: I1129 00:00:04.715033 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 00:00:04 crc kubenswrapper[4931]: I1129 00:00:04.718374 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:04 crc kubenswrapper[4931]: I1129 00:00:04.718456 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:04 crc kubenswrapper[4931]: I1129 00:00:04.718482 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:04 crc kubenswrapper[4931]: I1129 00:00:04.725161 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:00:04 crc kubenswrapper[4931]: I1129 00:00:04.858682 4931 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 29 00:00:04 crc kubenswrapper[4931]: I1129 00:00:04.858831 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 29 00:00:05 crc kubenswrapper[4931]: I1129 00:00:05.350512 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 00:00:05 crc kubenswrapper[4931]: I1129 00:00:05.352065 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:05 crc kubenswrapper[4931]: I1129 00:00:05.352137 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:05 crc kubenswrapper[4931]: I1129 00:00:05.352158 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:05 crc kubenswrapper[4931]: I1129 00:00:05.966483 4931 trace.go:236] Trace[2000243279]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 23:59:51.140) (total time: 14825ms): Nov 29 00:00:05 crc kubenswrapper[4931]: Trace[2000243279]: ---"Objects listed" error: 14825ms (00:00:05.966) Nov 29 00:00:05 crc kubenswrapper[4931]: Trace[2000243279]: [14.825834493s] [14.825834493s] END Nov 29 00:00:05 crc kubenswrapper[4931]: I1129 00:00:05.966528 4931 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 29 00:00:05 crc kubenswrapper[4931]: I1129 00:00:05.967799 4931 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 29 00:00:05 crc kubenswrapper[4931]: I1129 00:00:05.969626 4931 trace.go:236] Trace[229768910]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (28-Nov-2025 23:59:50.952) (total time: 15017ms): Nov 29 00:00:05 crc kubenswrapper[4931]: Trace[229768910]: ---"Objects listed" error: 15017ms (00:00:05.969) Nov 29 00:00:05 crc kubenswrapper[4931]: Trace[229768910]: [15.017479389s] [15.017479389s] END Nov 29 00:00:05 crc kubenswrapper[4931]: I1129 00:00:05.969681 4931 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 29 00:00:05 crc kubenswrapper[4931]: I1129 00:00:05.982350 4931 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.190250 4931 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42468->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.190309 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42468->192.168.126.11:17697: read: connection reset by peer" Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.190354 4931 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42478->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.190427 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42478->192.168.126.11:17697: read: connection reset by peer" Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.190712 4931 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.190836 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.191208 4931 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.191231 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.283631 4931 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.356472 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.360078 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e" exitCode=255 Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.360328 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e"} Nov 29 00:00:06 crc kubenswrapper[4931]: I1129 00:00:06.374882 4931 scope.go:117] "RemoveContainer" containerID="7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.129550 4931 apiserver.go:52] "Watching apiserver" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.134657 4931 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.135154 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.135704 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.135790 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.135840 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.135925 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.136202 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.136228 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.136242 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.136526 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.136591 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.141406 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.141707 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.141709 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.141869 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.141969 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.142095 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.142129 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.142957 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.143101 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.175369 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.193550 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.207004 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.219657 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.227983 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.232521 4931 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.233471 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.245470 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.264332 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276166 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276240 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276286 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276327 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276364 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276407 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276447 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276486 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276526 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276599 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276653 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276694 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276733 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276770 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276840 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276876 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276912 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276955 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.276989 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277024 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277060 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277059 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277312 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277327 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277366 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277329 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277279 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277547 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277595 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277633 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277668 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277705 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277745 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.277780 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280049 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280221 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280318 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280372 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280421 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280473 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280535 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280579 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280627 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280676 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280716 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280857 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280918 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.280970 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281027 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281067 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281118 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281204 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281249 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281298 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281348 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281392 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281451 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281499 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281592 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281652 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281790 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281878 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281869 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281896 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281963 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282180 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.281986 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282274 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282317 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282351 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282390 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282425 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282456 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282493 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282525 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282535 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282556 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282588 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282619 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282650 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282678 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282708 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282769 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282797 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282846 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282878 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282909 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282934 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282965 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282995 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283029 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283057 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283086 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283125 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283159 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283209 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283240 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283270 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283300 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283333 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283364 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283395 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283426 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283455 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283484 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283518 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283550 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283577 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283613 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283649 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283674 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283706 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283737 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283767 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283802 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284063 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284090 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284119 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284149 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284179 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284210 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284243 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284274 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284299 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284327 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284357 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284384 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284413 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284442 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284466 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284494 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284523 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284551 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284577 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284606 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284635 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284660 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284722 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.285781 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.285877 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.285924 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.285977 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286026 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286221 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286269 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286318 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286358 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286401 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286445 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286485 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286523 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286568 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286611 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286653 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286698 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286775 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286841 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286887 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286934 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.286978 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.287017 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.287061 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.287120 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.287162 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.287871 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.288221 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282543 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.282680 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283201 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.283355 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284126 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284656 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.284987 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.285466 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.285522 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.285959 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.292207 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.287214 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.287620 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.287720 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.297246 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.288017 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.287746 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.289117 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.292709 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.292744 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.292979 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.293599 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.288076 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.289572 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.294549 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.294645 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:00:07.794594839 +0000 UTC m=+20.956488111 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.294636 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.294920 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.291467 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.292157 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.296195 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.296183 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.296418 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.296602 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.296690 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.296467 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.297140 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.297192 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.297312 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.297338 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.297764 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.298192 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.298647 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.298867 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.298944 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.292518 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.299415 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.299481 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.299526 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.299565 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.299601 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.299639 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.299676 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.299712 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.299660 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.299774 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.299751 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.299845 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.293860 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300262 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300248 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300299 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300433 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300490 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300536 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300706 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300758 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300867 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300911 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300951 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300987 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301023 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301058 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301097 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301134 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301171 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301210 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301250 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301290 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301326 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301369 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301412 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301452 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301515 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301552 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301591 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301631 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301699 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301741 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301831 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301871 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302116 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302154 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302193 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302423 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302465 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302504 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302544 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302621 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302676 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302765 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302926 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302969 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303010 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303104 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303150 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303193 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303231 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303271 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303319 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303363 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303404 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304577 4931 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304666 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304740 4931 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304766 4931 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304883 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304906 4931 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304928 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304951 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304973 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304996 4931 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305016 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305037 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305059 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305079 4931 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305100 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305121 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305153 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305173 4931 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305195 4931 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305217 4931 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305244 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305264 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305288 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305313 4931 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305641 4931 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305674 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305696 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305759 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.308145 4931 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300794 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.300710 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301075 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301154 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301185 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301215 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.308846 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301312 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.295133 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301349 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301965 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.301871 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302428 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302622 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302656 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303044 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.302684 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303217 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303263 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303283 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303795 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303995 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303969 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304095 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304039 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304168 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.309286 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304415 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304695 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304762 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.309379 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.304961 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305279 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305302 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305344 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305497 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305693 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.305991 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.306159 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.306233 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.307084 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.307290 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.307594 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.307734 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.307755 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.307977 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.308765 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.308986 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.309227 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.303324 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.310079 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.310238 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.310385 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.310410 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.310624 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.310663 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.311186 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.311292 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.311303 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.311569 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.311589 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.312406 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.312443 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.312528 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.313349 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.313853 4931 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.313979 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.314108 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.314526 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.314784 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.314912 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.308612 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.315232 4931 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.315277 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.315314 4931 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.315345 4931 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.315379 4931 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.315415 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.315445 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.315475 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.316093 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.316591 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.316665 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.316866 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.317378 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.317739 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.318628 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.318970 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.319288 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:07.819253917 +0000 UTC m=+20.981147219 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.319361 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:07.81934419 +0000 UTC m=+20.981237682 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.319605 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.319836 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.319999 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.320233 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.320463 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.320727 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.321069 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.321104 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.325367 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.333969 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.334006 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.334024 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.334128 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:07.834099768 +0000 UTC m=+20.995993010 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.334170 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.337495 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.337555 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.337596 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.337738 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:07.837664532 +0000 UTC m=+20.999557774 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.337873 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.337927 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.338264 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.338388 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.342697 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.343036 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.343062 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.343255 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.344689 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.344712 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.344883 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.345515 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.345703 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.345763 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.346234 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.346272 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.346289 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.347129 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.347245 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.347505 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.347541 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.347889 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.348111 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.348426 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.348554 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.349138 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.349360 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.349626 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.350162 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.350535 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.350763 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.350981 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.351355 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.351769 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.352016 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.354087 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.354351 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.356260 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.356567 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.358331 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.358541 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.358635 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.358716 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.359241 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.359414 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.359534 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.359755 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.362457 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.365089 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.368987 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.371951 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f"} Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.372162 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.381330 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.393947 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.395686 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.397067 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.403953 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.405259 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.411217 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.416896 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.416984 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417038 4931 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417032 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417051 4931 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417115 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417130 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417145 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417156 4931 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417166 4931 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417176 4931 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417186 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417195 4931 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417205 4931 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417218 4931 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417228 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417237 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417246 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417255 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417264 4931 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417273 4931 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417325 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417336 4931 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417348 4931 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417359 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417371 4931 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417382 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417392 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417403 4931 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417411 4931 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417421 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417431 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417441 4931 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417452 4931 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417462 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417472 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417483 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417492 4931 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417504 4931 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417514 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417524 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417534 4931 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417544 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417554 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417563 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417572 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417582 4931 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417591 4931 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417602 4931 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417611 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417621 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417629 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417637 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417647 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417657 4931 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417668 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417678 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417687 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417695 4931 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417704 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417713 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417722 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417732 4931 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417741 4931 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417750 4931 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417761 4931 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417770 4931 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417779 4931 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417787 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417796 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417820 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417829 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417842 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417851 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417862 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417871 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417881 4931 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417889 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417899 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417907 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417916 4931 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417927 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417936 4931 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417944 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417953 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417962 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417973 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417984 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.417994 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418003 4931 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418012 4931 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418021 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418036 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418045 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418054 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418063 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418072 4931 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418082 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418091 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418099 4931 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418107 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418115 4931 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418123 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418132 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418141 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418150 4931 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418159 4931 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418167 4931 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418175 4931 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418184 4931 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418193 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418202 4931 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418211 4931 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418220 4931 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418230 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418239 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418247 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418256 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418265 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418275 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418283 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418292 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418300 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418309 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418317 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418325 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418333 4931 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418341 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418349 4931 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418359 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418367 4931 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418376 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418383 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418396 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418406 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418415 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418424 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418432 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418441 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418450 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418459 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418468 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418476 4931 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418485 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418493 4931 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418501 4931 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418509 4931 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418518 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418527 4931 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418537 4931 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418547 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418555 4931 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418567 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418576 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418585 4931 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418593 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418602 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418610 4931 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418620 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.418628 4931 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.422866 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.433050 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.442326 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.452895 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.462252 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.463333 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.472466 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.478499 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.480729 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.492613 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.503364 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.519830 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.535780 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.563199 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.838633 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.838720 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.838749 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.838774 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:07 crc kubenswrapper[4931]: I1129 00:00:07.838824 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.838915 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.838927 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:00:08.838886081 +0000 UTC m=+22.000779323 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.838978 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:08.838960823 +0000 UTC m=+22.000854055 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.839001 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.839044 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.839062 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.839074 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.839134 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:08.839110197 +0000 UTC m=+22.001003439 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.839159 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:08.839149958 +0000 UTC m=+22.001043200 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.839211 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.839228 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.839243 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:07 crc kubenswrapper[4931]: E1129 00:00:07.839276 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:08.839266121 +0000 UTC m=+22.001159373 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.211737 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.211904 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.375121 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e"} Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.375189 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0b8de375c4e678316340c5f34e58dd8e72b547f6fa5da00ef94bfd4665a59bc7"} Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.376225 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0a462f7914c526d9aa6811f99efff48ce0abea463636bb0ac9876208972ee809"} Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.378622 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b"} Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.378703 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba"} Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.378730 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"29e253a436133ab73d65083e61c0d47fb12a43833dab230874853808d1d13036"} Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.399642 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.420443 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.441928 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.466107 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.487198 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.506181 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.527628 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.544215 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.565890 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.583506 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.598218 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.614648 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.632149 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.651185 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.681251 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.705088 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:08Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.848163 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.848275 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.848325 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.848347 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:00:10.848317096 +0000 UTC m=+24.010210328 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.848380 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:08 crc kubenswrapper[4931]: I1129 00:00:08.848412 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.848426 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.848488 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:10.84847733 +0000 UTC m=+24.010370562 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.848563 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.848582 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.848597 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.848609 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.848687 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.848762 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.848789 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.848649 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:10.848634564 +0000 UTC m=+24.010527806 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.849058 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:10.848979563 +0000 UTC m=+24.010872835 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:08 crc kubenswrapper[4931]: E1129 00:00:08.849113 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:10.849098406 +0000 UTC m=+24.010991668 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.004367 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.017559 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.021249 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.029379 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.044419 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.066470 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.096517 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.113336 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.137244 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.154633 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.172499 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.189755 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.208954 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.211719 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.211842 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:09 crc kubenswrapper[4931]: E1129 00:00:09.211863 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:09 crc kubenswrapper[4931]: E1129 00:00:09.212118 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.216366 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.217251 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.218786 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.219635 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.220942 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.221641 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.222483 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.223751 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.224669 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.226172 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.226903 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.228542 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.229314 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.230035 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.231853 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.236838 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.237523 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.240255 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.240870 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.241635 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.243056 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.243659 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.244987 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.245571 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.247114 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.247769 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.248855 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.250315 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.250990 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.252434 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.255239 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.256654 4931 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.256935 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.261353 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.261635 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.263130 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.265398 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.278994 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.280401 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.282530 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.283906 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.284268 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.286205 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.287149 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.287978 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.289616 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.291072 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.291696 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.292959 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.293676 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.295190 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.295899 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.297282 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.297959 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.298689 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.302464 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.303229 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.305652 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.329717 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.345830 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: I1129 00:00:09.361656 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:09 crc kubenswrapper[4931]: E1129 00:00:09.391649 4931 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.212516 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.212801 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.261304 4931 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.263961 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.264016 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.264036 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.264130 4931 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.274228 4931 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.274369 4931 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.276219 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.276256 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.276273 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.276292 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.276313 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:10Z","lastTransitionTime":"2025-11-29T00:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.313551 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.319947 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.320018 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.320037 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.320064 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.320083 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:10Z","lastTransitionTime":"2025-11-29T00:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.341500 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.349984 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.350070 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.350093 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.350124 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.350145 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:10Z","lastTransitionTime":"2025-11-29T00:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.372347 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.377632 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.377687 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.377705 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.377729 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.377749 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:10Z","lastTransitionTime":"2025-11-29T00:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.398864 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.403596 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.403640 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.403657 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.403681 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.403698 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:10Z","lastTransitionTime":"2025-11-29T00:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.419971 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.420211 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.423440 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.423503 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.423522 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.423547 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.423564 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:10Z","lastTransitionTime":"2025-11-29T00:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.527540 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.527606 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.527630 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.527662 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.527683 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:10Z","lastTransitionTime":"2025-11-29T00:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.631681 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.631757 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.631776 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.631804 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.631852 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:10Z","lastTransitionTime":"2025-11-29T00:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.735081 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.735142 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.735160 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.735182 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.735198 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:10Z","lastTransitionTime":"2025-11-29T00:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.838927 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.838998 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.839019 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.839059 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.839081 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:10Z","lastTransitionTime":"2025-11-29T00:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.867704 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.867879 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.867938 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:00:14.867895045 +0000 UTC m=+28.029788287 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.868008 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.868073 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.868081 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.868100 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.868120 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.868138 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.868195 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:14.868170682 +0000 UTC m=+28.030063954 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.868234 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.868296 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:14.868277545 +0000 UTC m=+28.030170907 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.868343 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.868379 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.868427 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.868455 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.868489 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:14.868450109 +0000 UTC m=+28.030343421 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:10 crc kubenswrapper[4931]: E1129 00:00:10.868533 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:14.868508121 +0000 UTC m=+28.030401413 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.942451 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.942532 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.942552 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.942582 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:10 crc kubenswrapper[4931]: I1129 00:00:10.942603 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:10Z","lastTransitionTime":"2025-11-29T00:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.048455 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.048535 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.048559 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.048592 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.048616 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:11Z","lastTransitionTime":"2025-11-29T00:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.151763 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.151871 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.151898 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.151932 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.151971 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:11Z","lastTransitionTime":"2025-11-29T00:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.211694 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.211748 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:11 crc kubenswrapper[4931]: E1129 00:00:11.211856 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:11 crc kubenswrapper[4931]: E1129 00:00:11.211958 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.254464 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.254525 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.254535 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.254553 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.254566 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:11Z","lastTransitionTime":"2025-11-29T00:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.358205 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.358298 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.358310 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.358332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.358350 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:11Z","lastTransitionTime":"2025-11-29T00:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.461024 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.461204 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.461224 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.461256 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.461279 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:11Z","lastTransitionTime":"2025-11-29T00:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.563956 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.564022 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.564035 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.564054 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.564070 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:11Z","lastTransitionTime":"2025-11-29T00:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.667655 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.667706 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.667716 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.667734 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.667747 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:11Z","lastTransitionTime":"2025-11-29T00:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.771489 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.771556 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.771572 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.771599 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.771617 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:11Z","lastTransitionTime":"2025-11-29T00:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.862923 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.868174 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.878070 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.878153 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.878174 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.878207 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.878230 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:11Z","lastTransitionTime":"2025-11-29T00:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.884530 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.906005 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:11 crc kubenswrapper[4931]: I1129 00:00:11.939785 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:11.961776 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.007346 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.007407 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.007420 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.007467 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.007482 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:12Z","lastTransitionTime":"2025-11-29T00:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.020636 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.036293 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.056117 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.077183 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.099440 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.111474 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.111563 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.111586 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.111623 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.111649 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:12Z","lastTransitionTime":"2025-11-29T00:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.122130 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.152042 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.195169 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.212224 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:12 crc kubenswrapper[4931]: E1129 00:00:12.212485 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.214075 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.214148 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.214164 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.214184 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.214195 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:12Z","lastTransitionTime":"2025-11-29T00:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.215693 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.244304 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.267038 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.285206 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.304646 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.317249 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.317323 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.317342 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.317374 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.317396 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:12Z","lastTransitionTime":"2025-11-29T00:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.326430 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.393874 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15"} Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.419667 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.419741 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.419752 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.419774 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.419788 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:12Z","lastTransitionTime":"2025-11-29T00:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.433359 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.461473 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.486906 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.509702 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.523545 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.523622 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.523644 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.523673 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.523693 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:12Z","lastTransitionTime":"2025-11-29T00:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.529745 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.549056 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.566677 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.583365 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.599619 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.626918 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.626953 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.626962 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.627157 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.627169 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:12Z","lastTransitionTime":"2025-11-29T00:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.730628 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.730727 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.730857 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.730950 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.730971 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:12Z","lastTransitionTime":"2025-11-29T00:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.822697 4931 csr.go:261] certificate signing request csr-rz4c9 is approved, waiting to be issued Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.833611 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.833637 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.833646 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.833661 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.833680 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:12Z","lastTransitionTime":"2025-11-29T00:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.843466 4931 csr.go:257] certificate signing request csr-rz4c9 is issued Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.936007 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.936047 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.936099 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.936117 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:12 crc kubenswrapper[4931]: I1129 00:00:12.936127 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:12Z","lastTransitionTime":"2025-11-29T00:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.038902 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.038932 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.038940 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.038956 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.038966 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:13Z","lastTransitionTime":"2025-11-29T00:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.141418 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.141463 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.141476 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.141494 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.141507 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:13Z","lastTransitionTime":"2025-11-29T00:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.212091 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.212203 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:13 crc kubenswrapper[4931]: E1129 00:00:13.212265 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:13 crc kubenswrapper[4931]: E1129 00:00:13.212401 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.244175 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.244230 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.244243 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.244261 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.244274 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:13Z","lastTransitionTime":"2025-11-29T00:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.347422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.347546 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.347566 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.347593 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.347645 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:13Z","lastTransitionTime":"2025-11-29T00:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.450620 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.450674 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.450683 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.450705 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.450715 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:13Z","lastTransitionTime":"2025-11-29T00:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.553108 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.553160 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.553173 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.553191 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.553203 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:13Z","lastTransitionTime":"2025-11-29T00:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.656196 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.656249 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.656261 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.656284 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.656297 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:13Z","lastTransitionTime":"2025-11-29T00:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.758781 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.758845 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.758865 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.758881 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.758909 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:13Z","lastTransitionTime":"2025-11-29T00:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.844249 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-11-28 23:55:12 +0000 UTC, rotation deadline is 2026-09-26 04:44:36.277073353 +0000 UTC Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.844340 4931 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7228h44m22.432737708s for next certificate rotation Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.861890 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.861939 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.861949 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.861969 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.861982 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:13Z","lastTransitionTime":"2025-11-29T00:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.964669 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.964793 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.964850 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.964879 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:13 crc kubenswrapper[4931]: I1129 00:00:13.964937 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:13Z","lastTransitionTime":"2025-11-29T00:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.067989 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.068046 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.068061 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.068084 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.068096 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:14Z","lastTransitionTime":"2025-11-29T00:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.170633 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.170705 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.170719 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.170744 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.170759 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:14Z","lastTransitionTime":"2025-11-29T00:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.212043 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.212290 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.221699 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-6xxhj"] Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.222191 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.225966 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.226079 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-w99nx"] Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.226364 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.226767 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-w99nx" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.228857 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.228891 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b227491-f635-4817-96ed-90b96ab6aab3-mcd-auth-proxy-config\") pod \"machine-config-daemon-6xxhj\" (UID: \"0b227491-f635-4817-96ed-90b96ab6aab3\") " pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.229051 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0b227491-f635-4817-96ed-90b96ab6aab3-rootfs\") pod \"machine-config-daemon-6xxhj\" (UID: \"0b227491-f635-4817-96ed-90b96ab6aab3\") " pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.229274 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b227491-f635-4817-96ed-90b96ab6aab3-proxy-tls\") pod \"machine-config-daemon-6xxhj\" (UID: \"0b227491-f635-4817-96ed-90b96ab6aab3\") " pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.229378 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.229447 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hhsp\" (UniqueName: \"kubernetes.io/projected/0b227491-f635-4817-96ed-90b96ab6aab3-kube-api-access-7hhsp\") pod \"machine-config-daemon-6xxhj\" (UID: \"0b227491-f635-4817-96ed-90b96ab6aab3\") " pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.229679 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.229863 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.230131 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.231388 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.244527 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.264960 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.274403 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.274478 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.274503 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.274542 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.274575 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:14Z","lastTransitionTime":"2025-11-29T00:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.287525 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.305912 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.324969 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.330161 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjjkv\" (UniqueName: \"kubernetes.io/projected/3f08fd4c-2846-497a-955d-c66ba5b58a9c-kube-api-access-vjjkv\") pod \"node-resolver-w99nx\" (UID: \"3f08fd4c-2846-497a-955d-c66ba5b58a9c\") " pod="openshift-dns/node-resolver-w99nx" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.330217 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b227491-f635-4817-96ed-90b96ab6aab3-mcd-auth-proxy-config\") pod \"machine-config-daemon-6xxhj\" (UID: \"0b227491-f635-4817-96ed-90b96ab6aab3\") " pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.330463 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0b227491-f635-4817-96ed-90b96ab6aab3-rootfs\") pod \"machine-config-daemon-6xxhj\" (UID: \"0b227491-f635-4817-96ed-90b96ab6aab3\") " pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.330959 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b227491-f635-4817-96ed-90b96ab6aab3-mcd-auth-proxy-config\") pod \"machine-config-daemon-6xxhj\" (UID: \"0b227491-f635-4817-96ed-90b96ab6aab3\") " pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.331021 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0b227491-f635-4817-96ed-90b96ab6aab3-rootfs\") pod \"machine-config-daemon-6xxhj\" (UID: \"0b227491-f635-4817-96ed-90b96ab6aab3\") " pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.331057 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3f08fd4c-2846-497a-955d-c66ba5b58a9c-hosts-file\") pod \"node-resolver-w99nx\" (UID: \"3f08fd4c-2846-497a-955d-c66ba5b58a9c\") " pod="openshift-dns/node-resolver-w99nx" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.331110 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b227491-f635-4817-96ed-90b96ab6aab3-proxy-tls\") pod \"machine-config-daemon-6xxhj\" (UID: \"0b227491-f635-4817-96ed-90b96ab6aab3\") " pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.331773 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hhsp\" (UniqueName: \"kubernetes.io/projected/0b227491-f635-4817-96ed-90b96ab6aab3-kube-api-access-7hhsp\") pod \"machine-config-daemon-6xxhj\" (UID: \"0b227491-f635-4817-96ed-90b96ab6aab3\") " pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.349223 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b227491-f635-4817-96ed-90b96ab6aab3-proxy-tls\") pod \"machine-config-daemon-6xxhj\" (UID: \"0b227491-f635-4817-96ed-90b96ab6aab3\") " pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.359988 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hhsp\" (UniqueName: \"kubernetes.io/projected/0b227491-f635-4817-96ed-90b96ab6aab3-kube-api-access-7hhsp\") pod \"machine-config-daemon-6xxhj\" (UID: \"0b227491-f635-4817-96ed-90b96ab6aab3\") " pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.373323 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.377404 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.377439 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.377452 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.377488 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.377502 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:14Z","lastTransitionTime":"2025-11-29T00:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.396524 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.417468 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.432762 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjjkv\" (UniqueName: \"kubernetes.io/projected/3f08fd4c-2846-497a-955d-c66ba5b58a9c-kube-api-access-vjjkv\") pod \"node-resolver-w99nx\" (UID: \"3f08fd4c-2846-497a-955d-c66ba5b58a9c\") " pod="openshift-dns/node-resolver-w99nx" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.432854 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3f08fd4c-2846-497a-955d-c66ba5b58a9c-hosts-file\") pod \"node-resolver-w99nx\" (UID: \"3f08fd4c-2846-497a-955d-c66ba5b58a9c\") " pod="openshift-dns/node-resolver-w99nx" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.432958 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3f08fd4c-2846-497a-955d-c66ba5b58a9c-hosts-file\") pod \"node-resolver-w99nx\" (UID: \"3f08fd4c-2846-497a-955d-c66ba5b58a9c\") " pod="openshift-dns/node-resolver-w99nx" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.434685 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.450490 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjjkv\" (UniqueName: \"kubernetes.io/projected/3f08fd4c-2846-497a-955d-c66ba5b58a9c-kube-api-access-vjjkv\") pod \"node-resolver-w99nx\" (UID: \"3f08fd4c-2846-497a-955d-c66ba5b58a9c\") " pod="openshift-dns/node-resolver-w99nx" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.453014 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.466731 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.480877 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.480928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.480940 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.480961 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.480976 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:14Z","lastTransitionTime":"2025-11-29T00:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.481499 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.501281 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.517179 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.534471 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.540323 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.545515 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-w99nx" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.552268 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: W1129 00:00:14.562824 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f08fd4c_2846_497a_955d_c66ba5b58a9c.slice/crio-27b79df468a95836878bd0b941a12b2e0e6e77978a09fd21e9e406e64b4e2b36 WatchSource:0}: Error finding container 27b79df468a95836878bd0b941a12b2e0e6e77978a09fd21e9e406e64b4e2b36: Status 404 returned error can't find the container with id 27b79df468a95836878bd0b941a12b2e0e6e77978a09fd21e9e406e64b4e2b36 Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.569227 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.584502 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.584622 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.584636 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.584654 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.585127 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:14Z","lastTransitionTime":"2025-11-29T00:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.587019 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.606983 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.619707 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-n5l7h"] Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.622417 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.626721 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-22vkg"] Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.627927 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.629693 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.630357 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.630381 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.630831 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.631218 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.631512 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.631802 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.635234 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-system-cni-dir\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.635307 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-tuning-conf-dir\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.635340 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-cni-binary-copy\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.635388 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.635415 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-os-release\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.635568 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-cnibin\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.635618 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szjx9\" (UniqueName: \"kubernetes.io/projected/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-kube-api-access-szjx9\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.639838 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.656838 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.679108 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.688881 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.688941 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.688955 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.688976 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.688989 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:14Z","lastTransitionTime":"2025-11-29T00:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.695030 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.707585 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.719009 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.734638 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.736939 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-os-release\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737138 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-os-release\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737081 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-os-release\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737229 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-run-multus-certs\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737301 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xl2w\" (UniqueName: \"kubernetes.io/projected/31e6835b-462a-4a4f-9ba9-74531c14d41a-kube-api-access-2xl2w\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737363 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-system-cni-dir\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737390 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/31e6835b-462a-4a4f-9ba9-74531c14d41a-cni-binary-copy\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737456 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-var-lib-cni-multus\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737491 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-cnibin\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737570 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-cnibin\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737515 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szjx9\" (UniqueName: \"kubernetes.io/projected/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-kube-api-access-szjx9\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737627 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-run-k8s-cni-cncf-io\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737657 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-run-netns\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737700 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-cnibin\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737725 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-var-lib-cni-bin\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737748 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-tuning-conf-dir\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737772 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-hostroot\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737800 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/31e6835b-462a-4a4f-9ba9-74531c14d41a-multus-daemon-config\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737846 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737894 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-multus-socket-dir-parent\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737920 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-var-lib-kubelet\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737947 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-system-cni-dir\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.737975 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-etc-kubernetes\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.738002 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-multus-cni-dir\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.738025 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-cni-binary-copy\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.738047 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-multus-conf-dir\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.738190 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-system-cni-dir\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.738380 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-tuning-conf-dir\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.738969 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.740320 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-cni-binary-copy\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.750631 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.758443 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szjx9\" (UniqueName: \"kubernetes.io/projected/bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde-kube-api-access-szjx9\") pod \"multus-additional-cni-plugins-n5l7h\" (UID: \"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\") " pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.764530 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.789620 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.793041 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.793091 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.793109 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.793133 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.793149 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:14Z","lastTransitionTime":"2025-11-29T00:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.805084 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.835671 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839050 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-multus-conf-dir\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839102 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-system-cni-dir\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839128 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-os-release\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839153 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-run-multus-certs\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839205 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xl2w\" (UniqueName: \"kubernetes.io/projected/31e6835b-462a-4a4f-9ba9-74531c14d41a-kube-api-access-2xl2w\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839229 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/31e6835b-462a-4a4f-9ba9-74531c14d41a-cni-binary-copy\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839253 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-var-lib-cni-multus\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839278 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-run-netns\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839299 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-run-k8s-cni-cncf-io\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839329 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-cnibin\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839353 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-var-lib-cni-bin\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839384 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-hostroot\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839409 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/31e6835b-462a-4a4f-9ba9-74531c14d41a-multus-daemon-config\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839455 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-multus-socket-dir-parent\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839451 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-system-cni-dir\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839487 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-var-lib-kubelet\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839481 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-multus-conf-dir\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839525 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-var-lib-cni-multus\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839514 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-etc-kubernetes\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839588 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-run-netns\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839551 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-run-multus-certs\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839562 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-etc-kubernetes\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839616 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-multus-cni-dir\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839686 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-run-k8s-cni-cncf-io\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839742 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-cnibin\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839774 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-var-lib-cni-bin\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839784 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-os-release\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839831 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-hostroot\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839871 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-host-var-lib-kubelet\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839980 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-multus-socket-dir-parent\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.839983 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/31e6835b-462a-4a4f-9ba9-74531c14d41a-multus-cni-dir\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.840605 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/31e6835b-462a-4a4f-9ba9-74531c14d41a-cni-binary-copy\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.840770 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/31e6835b-462a-4a4f-9ba9-74531c14d41a-multus-daemon-config\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.852529 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.860189 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xl2w\" (UniqueName: \"kubernetes.io/projected/31e6835b-462a-4a4f-9ba9-74531c14d41a-kube-api-access-2xl2w\") pod \"multus-22vkg\" (UID: \"31e6835b-462a-4a4f-9ba9-74531c14d41a\") " pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.870323 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.890559 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:14Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.895576 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.895647 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.895667 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.895694 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.895714 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:14Z","lastTransitionTime":"2025-11-29T00:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.940261 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940401 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:00:22.940371806 +0000 UTC m=+36.102265048 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.940464 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.940514 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.940544 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.940572 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940614 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940672 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:22.940661574 +0000 UTC m=+36.102554806 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940687 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940705 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940725 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940728 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940789 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940848 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940867 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940777 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:22.940764146 +0000 UTC m=+36.102657398 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940936 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:22.94091597 +0000 UTC m=+36.102809202 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:14 crc kubenswrapper[4931]: E1129 00:00:14.940958 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:22.940951621 +0000 UTC m=+36.102844853 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.961413 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" Nov 29 00:00:14 crc kubenswrapper[4931]: I1129 00:00:14.970545 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-22vkg" Nov 29 00:00:14 crc kubenswrapper[4931]: W1129 00:00:14.978188 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc6c5917_e8b2_45a3_95a6_04ccb7ae8fde.slice/crio-07cfcd826914eb0534bf7588d90ac1bc175938ec4243b0204854caefe3b738bc WatchSource:0}: Error finding container 07cfcd826914eb0534bf7588d90ac1bc175938ec4243b0204854caefe3b738bc: Status 404 returned error can't find the container with id 07cfcd826914eb0534bf7588d90ac1bc175938ec4243b0204854caefe3b738bc Nov 29 00:00:14 crc kubenswrapper[4931]: W1129 00:00:14.989778 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31e6835b_462a_4a4f_9ba9_74531c14d41a.slice/crio-c664844df4146f1cf4ac55cece6609530c91e0ccc94a25827e0a3a61099e6234 WatchSource:0}: Error finding container c664844df4146f1cf4ac55cece6609530c91e0ccc94a25827e0a3a61099e6234: Status 404 returned error can't find the container with id c664844df4146f1cf4ac55cece6609530c91e0ccc94a25827e0a3a61099e6234 Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.000880 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4m4l4"] Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.001848 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.001887 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.001902 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.001919 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.001931 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:15Z","lastTransitionTime":"2025-11-29T00:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.001985 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.005866 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.005938 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.006282 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.006621 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.006804 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.009321 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.009567 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.026119 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.041367 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-slash\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.041436 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-env-overrides\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.041508 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-kubelet\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.041543 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-openvswitch\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.041596 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-ovn\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.041635 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.041728 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfsrg\" (UniqueName: \"kubernetes.io/projected/2fae9705-fac9-4c99-b5f8-280e5c05e450-kube-api-access-nfsrg\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.041784 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-log-socket\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.041863 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-cni-bin\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.042055 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-systemd\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.042108 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-var-lib-openvswitch\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.042147 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovnkube-script-lib\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.042182 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-run-ovn-kubernetes\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.042207 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-cni-netd\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.042278 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovnkube-config\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.042358 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-node-log\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.042431 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-systemd-units\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.042459 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-etc-openvswitch\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.042493 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovn-node-metrics-cert\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.042545 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-run-netns\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.044246 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.059947 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.075437 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.090498 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.106887 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.108663 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.108734 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.108753 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.108780 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.108800 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:15Z","lastTransitionTime":"2025-11-29T00:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.122588 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.138186 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.143846 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-kubelet\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.143892 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-openvswitch\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.143928 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-ovn\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.143954 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144019 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfsrg\" (UniqueName: \"kubernetes.io/projected/2fae9705-fac9-4c99-b5f8-280e5c05e450-kube-api-access-nfsrg\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144044 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-log-socket\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144066 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-cni-bin\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144090 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-systemd\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144109 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-var-lib-openvswitch\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144130 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovnkube-script-lib\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144155 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-run-ovn-kubernetes\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144176 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-cni-netd\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144195 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovnkube-config\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144215 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-node-log\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144238 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-systemd-units\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144266 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-etc-openvswitch\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144289 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovn-node-metrics-cert\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144314 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-run-netns\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144339 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-slash\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.144367 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-env-overrides\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.145189 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-env-overrides\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.145267 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-cni-bin\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.145518 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-slash\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.145546 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-ovn\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.145752 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-node-log\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.145999 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-etc-openvswitch\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.146033 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-systemd\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.146062 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.146087 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-var-lib-openvswitch\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.146113 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-run-netns\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.146951 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-systemd-units\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.146973 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-log-socket\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.146984 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-openvswitch\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.146995 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-kubelet\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.147067 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-cni-netd\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.147113 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-run-ovn-kubernetes\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.147192 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovnkube-script-lib\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.147588 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovnkube-config\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.154672 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovn-node-metrics-cert\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.156033 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.168548 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfsrg\" (UniqueName: \"kubernetes.io/projected/2fae9705-fac9-4c99-b5f8-280e5c05e450-kube-api-access-nfsrg\") pod \"ovnkube-node-4m4l4\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.186298 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.211232 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:15 crc kubenswrapper[4931]: E1129 00:00:15.211369 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.214058 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:15 crc kubenswrapper[4931]: E1129 00:00:15.214493 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.215104 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.215159 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.215177 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.215204 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.215225 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:15Z","lastTransitionTime":"2025-11-29T00:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.218850 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.232661 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.246379 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.256753 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.318520 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.318563 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.318573 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.318592 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.318617 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:15Z","lastTransitionTime":"2025-11-29T00:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.324348 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:15 crc kubenswrapper[4931]: W1129 00:00:15.336045 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fae9705_fac9_4c99_b5f8_280e5c05e450.slice/crio-1e64dbe7efef590a70b3fa79084f97cb036c5a7dbd9bca3bebb5759f774a2662 WatchSource:0}: Error finding container 1e64dbe7efef590a70b3fa79084f97cb036c5a7dbd9bca3bebb5759f774a2662: Status 404 returned error can't find the container with id 1e64dbe7efef590a70b3fa79084f97cb036c5a7dbd9bca3bebb5759f774a2662 Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.404506 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22vkg" event={"ID":"31e6835b-462a-4a4f-9ba9-74531c14d41a","Type":"ContainerStarted","Data":"c664844df4146f1cf4ac55cece6609530c91e0ccc94a25827e0a3a61099e6234"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.407128 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.407171 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.407191 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"773d478613c9dafc0ebad556cd53be1dae81d4b805ffc38bac83130a9b06db5e"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.418354 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"1e64dbe7efef590a70b3fa79084f97cb036c5a7dbd9bca3bebb5759f774a2662"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.423111 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.423151 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.423163 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.423184 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.423197 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:15Z","lastTransitionTime":"2025-11-29T00:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.423760 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" event={"ID":"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde","Type":"ContainerStarted","Data":"07cfcd826914eb0534bf7588d90ac1bc175938ec4243b0204854caefe3b738bc"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.426160 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-w99nx" event={"ID":"3f08fd4c-2846-497a-955d-c66ba5b58a9c","Type":"ContainerStarted","Data":"be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.426241 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-w99nx" event={"ID":"3f08fd4c-2846-497a-955d-c66ba5b58a9c","Type":"ContainerStarted","Data":"27b79df468a95836878bd0b941a12b2e0e6e77978a09fd21e9e406e64b4e2b36"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.428410 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.443775 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.457076 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.476470 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.489590 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.503448 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.518884 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.528446 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.528497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.528511 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.528529 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.528542 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:15Z","lastTransitionTime":"2025-11-29T00:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.534021 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.555726 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.603017 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.621473 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.631304 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.631353 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.631363 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.631380 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.631395 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:15Z","lastTransitionTime":"2025-11-29T00:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.646223 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.665544 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.684206 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.699253 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.711802 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.723264 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.734194 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.734232 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.734242 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.734259 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.734270 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:15Z","lastTransitionTime":"2025-11-29T00:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.747174 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.761627 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.775240 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.787315 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.799932 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.817188 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.837393 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.840773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.840823 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.840834 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.840851 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.840862 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:15Z","lastTransitionTime":"2025-11-29T00:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.860641 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.876699 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.891270 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.906499 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:15Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.944371 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.944435 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.944449 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.944470 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:15 crc kubenswrapper[4931]: I1129 00:00:15.944484 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:15Z","lastTransitionTime":"2025-11-29T00:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.047618 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.048111 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.048173 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.048234 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.048289 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:16Z","lastTransitionTime":"2025-11-29T00:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.151607 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.151652 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.151664 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.151681 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.151695 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:16Z","lastTransitionTime":"2025-11-29T00:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.212049 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:16 crc kubenswrapper[4931]: E1129 00:00:16.212463 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.255024 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.255071 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.255085 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.255109 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.255127 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:16Z","lastTransitionTime":"2025-11-29T00:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.357995 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.358502 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.358712 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.358899 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.359078 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:16Z","lastTransitionTime":"2025-11-29T00:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.432264 4931 generic.go:334] "Generic (PLEG): container finished" podID="bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde" containerID="5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69" exitCode=0 Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.432374 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" event={"ID":"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde","Type":"ContainerDied","Data":"5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.434774 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e" exitCode=0 Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.434846 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.436704 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22vkg" event={"ID":"31e6835b-462a-4a4f-9ba9-74531c14d41a","Type":"ContainerStarted","Data":"eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.467684 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.468995 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.469092 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.469206 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.469301 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:16Z","lastTransitionTime":"2025-11-29T00:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.470625 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.491380 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.504920 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.515572 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-n2tnq"] Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.516867 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-n2tnq" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.520844 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.521325 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.521477 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.521687 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.533382 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.562942 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.565957 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzcx6\" (UniqueName: \"kubernetes.io/projected/c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd-kube-api-access-tzcx6\") pod \"node-ca-n2tnq\" (UID: \"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\") " pod="openshift-image-registry/node-ca-n2tnq" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.566073 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd-host\") pod \"node-ca-n2tnq\" (UID: \"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\") " pod="openshift-image-registry/node-ca-n2tnq" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.566098 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd-serviceca\") pod \"node-ca-n2tnq\" (UID: \"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\") " pod="openshift-image-registry/node-ca-n2tnq" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.572783 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.572835 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.572847 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.572865 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.572877 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:16Z","lastTransitionTime":"2025-11-29T00:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.593059 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.653971 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.668018 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzcx6\" (UniqueName: \"kubernetes.io/projected/c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd-kube-api-access-tzcx6\") pod \"node-ca-n2tnq\" (UID: \"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\") " pod="openshift-image-registry/node-ca-n2tnq" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.668106 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd-host\") pod \"node-ca-n2tnq\" (UID: \"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\") " pod="openshift-image-registry/node-ca-n2tnq" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.668126 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd-serviceca\") pod \"node-ca-n2tnq\" (UID: \"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\") " pod="openshift-image-registry/node-ca-n2tnq" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.668649 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd-host\") pod \"node-ca-n2tnq\" (UID: \"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\") " pod="openshift-image-registry/node-ca-n2tnq" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.669244 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd-serviceca\") pod \"node-ca-n2tnq\" (UID: \"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\") " pod="openshift-image-registry/node-ca-n2tnq" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.673995 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.688380 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.688435 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.688447 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.688467 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.688480 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:16Z","lastTransitionTime":"2025-11-29T00:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.693092 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.702037 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzcx6\" (UniqueName: \"kubernetes.io/projected/c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd-kube-api-access-tzcx6\") pod \"node-ca-n2tnq\" (UID: \"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\") " pod="openshift-image-registry/node-ca-n2tnq" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.709084 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.727996 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.745840 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.769247 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.785922 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.791349 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.791404 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.791417 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.791438 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.791451 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:16Z","lastTransitionTime":"2025-11-29T00:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.803092 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.819009 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.838445 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.850934 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-n2tnq" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.859249 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: W1129 00:00:16.864946 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2ac4c77_c7aa_40ec_8bd9_e6afa4b2a8cd.slice/crio-8b2ba70243cfbaf9691cbba37d62eccd0bb18d360a96769b384f93169f123397 WatchSource:0}: Error finding container 8b2ba70243cfbaf9691cbba37d62eccd0bb18d360a96769b384f93169f123397: Status 404 returned error can't find the container with id 8b2ba70243cfbaf9691cbba37d62eccd0bb18d360a96769b384f93169f123397 Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.878944 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.895131 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.895836 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.895907 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.895924 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.895948 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.895962 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:16Z","lastTransitionTime":"2025-11-29T00:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.916058 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.946269 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.960612 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.978700 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.993610 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:16Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:16 crc kubenswrapper[4931]: I1129 00:00:16.995233 4931 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Nov 29 00:00:16 crc kubenswrapper[4931]: W1129 00:00:16.997437 4931 reflector.go:484] object-"openshift-image-registry"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Nov 29 00:00:17 crc kubenswrapper[4931]: W1129 00:00:16.999080 4931 reflector.go:484] object-"openshift-image-registry"/"image-registry-certificates": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"image-registry-certificates": Unexpected watch close - watch lasted less than a second and no items received Nov 29 00:00:17 crc kubenswrapper[4931]: W1129 00:00:16.999716 4931 reflector.go:484] object-"openshift-image-registry"/"node-ca-dockercfg-4777p": watch of *v1.Secret ended with: very short watch: object-"openshift-image-registry"/"node-ca-dockercfg-4777p": Unexpected watch close - watch lasted less than a second and no items received Nov 29 00:00:17 crc kubenswrapper[4931]: W1129 00:00:16.999778 4931 reflector.go:484] object-"openshift-image-registry"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.012762 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.012867 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.012886 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.012914 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.012934 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:17Z","lastTransitionTime":"2025-11-29T00:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.116629 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.116675 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.116687 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.116707 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.116720 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:17Z","lastTransitionTime":"2025-11-29T00:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.211418 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:17 crc kubenswrapper[4931]: E1129 00:00:17.211564 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.211838 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:17 crc kubenswrapper[4931]: E1129 00:00:17.212051 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.218890 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.218954 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.218975 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.219004 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.219027 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:17Z","lastTransitionTime":"2025-11-29T00:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.321468 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.321524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.321537 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.321557 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.321571 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:17Z","lastTransitionTime":"2025-11-29T00:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.425023 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.425413 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.425424 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.425441 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.425454 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:17Z","lastTransitionTime":"2025-11-29T00:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.448168 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.448230 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.448249 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.448266 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.448283 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.448299 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.451391 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-n2tnq" event={"ID":"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd","Type":"ContainerStarted","Data":"a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.451419 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-n2tnq" event={"ID":"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd","Type":"ContainerStarted","Data":"8b2ba70243cfbaf9691cbba37d62eccd0bb18d360a96769b384f93169f123397"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.454251 4931 generic.go:334] "Generic (PLEG): container finished" podID="bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde" containerID="a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412" exitCode=0 Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.454930 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" event={"ID":"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde","Type":"ContainerDied","Data":"a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.530656 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.530735 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.530759 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.530831 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.530858 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:17Z","lastTransitionTime":"2025-11-29T00:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.633694 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.633775 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.633802 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.633910 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.633936 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:17Z","lastTransitionTime":"2025-11-29T00:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.737574 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.737635 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.737655 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.737701 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.737721 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:17Z","lastTransitionTime":"2025-11-29T00:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.840847 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.840921 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.840946 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.840974 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.840992 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:17Z","lastTransitionTime":"2025-11-29T00:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.944105 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.944152 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.944164 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.944184 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:17 crc kubenswrapper[4931]: I1129 00:00:17.944197 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:17Z","lastTransitionTime":"2025-11-29T00:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.028392 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.047349 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.047425 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.047444 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.047474 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.047495 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:18Z","lastTransitionTime":"2025-11-29T00:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.073115 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.094511 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.111773 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.128380 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.130044 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.147198 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.151084 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.151167 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.151194 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.151232 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.151260 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:18Z","lastTransitionTime":"2025-11-29T00:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.156126 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.165822 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.180113 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.182428 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.189653 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.199116 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.211766 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:18 crc kubenswrapper[4931]: E1129 00:00:18.211990 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.217879 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.240152 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.255695 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.255759 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.255785 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.255858 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.255890 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:18Z","lastTransitionTime":"2025-11-29T00:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.258176 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.292564 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.327054 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.349913 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.359768 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.359846 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.359865 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.359895 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.359913 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:18Z","lastTransitionTime":"2025-11-29T00:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.371310 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.387487 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.404965 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.424631 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.445499 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.460875 4931 generic.go:334] "Generic (PLEG): container finished" podID="bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde" containerID="625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce" exitCode=0 Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.460928 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" event={"ID":"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde","Type":"ContainerDied","Data":"625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce"} Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.462079 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.462109 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.462121 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.462136 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.462149 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:18Z","lastTransitionTime":"2025-11-29T00:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.462833 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.487467 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.507285 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.520209 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.542242 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.586007 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.586050 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.586063 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.586084 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.586099 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:18Z","lastTransitionTime":"2025-11-29T00:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.586227 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.638297 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.659329 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.672893 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.687267 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.694250 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.694355 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.694381 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.694654 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.694768 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:18Z","lastTransitionTime":"2025-11-29T00:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.706173 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.716975 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.731491 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.747298 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.761992 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.775555 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.791601 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.797910 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.797946 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.797956 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.797973 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.797988 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:18Z","lastTransitionTime":"2025-11-29T00:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.805909 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.826357 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.842550 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.858683 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.889870 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.900181 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.900232 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.900250 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.900276 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.900296 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:18Z","lastTransitionTime":"2025-11-29T00:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.909157 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.930440 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.946388 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.965369 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.984406 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:18 crc kubenswrapper[4931]: I1129 00:00:18.998446 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:18Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.002749 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.002795 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.002828 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.002848 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.002862 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:19Z","lastTransitionTime":"2025-11-29T00:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.013241 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.106309 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.106384 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.106397 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.106420 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.106434 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:19Z","lastTransitionTime":"2025-11-29T00:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.209799 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.209863 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.209875 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.209892 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.209904 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:19Z","lastTransitionTime":"2025-11-29T00:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.225692 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:19 crc kubenswrapper[4931]: E1129 00:00:19.225910 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.225895 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:19 crc kubenswrapper[4931]: E1129 00:00:19.226945 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.312935 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.313002 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.313030 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.313097 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.313123 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:19Z","lastTransitionTime":"2025-11-29T00:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.417078 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.417151 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.417168 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.417196 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.417217 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:19Z","lastTransitionTime":"2025-11-29T00:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.470265 4931 generic.go:334] "Generic (PLEG): container finished" podID="bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde" containerID="c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724" exitCode=0 Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.470379 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" event={"ID":"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde","Type":"ContainerDied","Data":"c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724"} Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.479007 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0"} Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.499190 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.520990 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.521055 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.521073 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.521102 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.521121 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:19Z","lastTransitionTime":"2025-11-29T00:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.526376 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.554581 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.576324 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.592621 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.607755 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.624728 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.624788 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.624833 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.624861 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.624881 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:19Z","lastTransitionTime":"2025-11-29T00:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.629908 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.646487 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.662914 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.682365 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.701685 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.724470 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.727776 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.727846 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.727860 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.727879 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.727891 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:19Z","lastTransitionTime":"2025-11-29T00:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.751294 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.767036 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.781118 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:19Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.831924 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.831972 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.831986 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.832009 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.832024 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:19Z","lastTransitionTime":"2025-11-29T00:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.935314 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.935370 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.935388 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.935414 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:19 crc kubenswrapper[4931]: I1129 00:00:19.935435 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:19Z","lastTransitionTime":"2025-11-29T00:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.041955 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.042009 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.042026 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.042052 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.042069 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.145137 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.145996 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.146413 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.146879 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.147332 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.212407 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:20 crc kubenswrapper[4931]: E1129 00:00:20.213081 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.251110 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.251451 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.251590 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.251716 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.251874 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.355089 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.355466 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.355608 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.355738 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.355922 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.459531 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.459612 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.459637 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.459672 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.459695 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.489562 4931 generic.go:334] "Generic (PLEG): container finished" podID="bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde" containerID="de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1" exitCode=0 Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.489631 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" event={"ID":"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde","Type":"ContainerDied","Data":"de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1"} Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.513671 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.534099 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.561272 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.563091 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.563127 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.563138 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.563158 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.563172 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.577770 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.590024 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.612143 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.631665 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.658706 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.672756 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.673990 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.674207 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.674237 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.674488 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.701640 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.718232 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.737424 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.757173 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.774542 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.777388 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.777484 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.777540 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.777600 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.777659 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.797220 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.801341 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.801418 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.801438 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.801465 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.801485 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.814230 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: E1129 00:00:20.821537 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.827524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.827581 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.827600 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.827628 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.827651 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: E1129 00:00:20.849341 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.854771 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.854877 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.854897 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.854925 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.854943 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: E1129 00:00:20.876584 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.882801 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.882867 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.882881 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.882902 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.882917 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: E1129 00:00:20.903107 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.909007 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.909081 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.909102 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.909136 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.909170 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:20 crc kubenswrapper[4931]: E1129 00:00:20.932433 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:20Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:20 crc kubenswrapper[4931]: E1129 00:00:20.932662 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.935345 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.935409 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.935427 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.935455 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:20 crc kubenswrapper[4931]: I1129 00:00:20.935475 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:20Z","lastTransitionTime":"2025-11-29T00:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.038977 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.039038 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.039058 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.039086 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.039106 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:21Z","lastTransitionTime":"2025-11-29T00:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.142701 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.142765 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.142783 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.142842 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.142873 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:21Z","lastTransitionTime":"2025-11-29T00:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.211663 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:21 crc kubenswrapper[4931]: E1129 00:00:21.211900 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.212255 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:21 crc kubenswrapper[4931]: E1129 00:00:21.212592 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.246660 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.246729 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.246750 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.246784 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.246802 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:21Z","lastTransitionTime":"2025-11-29T00:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.350847 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.350904 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.350923 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.350951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.350969 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:21Z","lastTransitionTime":"2025-11-29T00:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.454751 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.454851 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.454868 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.454897 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.454917 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:21Z","lastTransitionTime":"2025-11-29T00:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.500318 4931 generic.go:334] "Generic (PLEG): container finished" podID="bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde" containerID="40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e" exitCode=0 Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.500392 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" event={"ID":"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde","Type":"ContainerDied","Data":"40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e"} Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.524282 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.548312 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.560078 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.560359 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.560508 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.560698 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.560867 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:21Z","lastTransitionTime":"2025-11-29T00:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.573687 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.596164 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.610682 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.624358 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.645871 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.663874 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.663921 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.663958 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.663979 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.663991 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:21Z","lastTransitionTime":"2025-11-29T00:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.665468 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.680138 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.702589 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.722776 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.753254 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.767151 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.767205 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.767218 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.767241 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.767256 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:21Z","lastTransitionTime":"2025-11-29T00:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.781525 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.800855 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.822491 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:21Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.870606 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.870652 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.870663 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.870682 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.870695 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:21Z","lastTransitionTime":"2025-11-29T00:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.973870 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.973917 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.973929 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.973947 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:21 crc kubenswrapper[4931]: I1129 00:00:21.973962 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:21Z","lastTransitionTime":"2025-11-29T00:00:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.077833 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.077875 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.077885 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.077914 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.077927 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:22Z","lastTransitionTime":"2025-11-29T00:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.181098 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.181151 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.181170 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.181196 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.181214 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:22Z","lastTransitionTime":"2025-11-29T00:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.211726 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.211966 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.283881 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.283941 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.283955 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.283977 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.283995 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:22Z","lastTransitionTime":"2025-11-29T00:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.387405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.387478 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.387497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.387525 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.387551 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:22Z","lastTransitionTime":"2025-11-29T00:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.490545 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.490598 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.490615 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.490640 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.490659 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:22Z","lastTransitionTime":"2025-11-29T00:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.513267 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba"} Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.513571 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.514054 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.514138 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.520644 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" event={"ID":"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde","Type":"ContainerStarted","Data":"e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7"} Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.540727 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.587069 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.589424 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.589694 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.594019 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.594417 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.594610 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.594849 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.595059 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:22Z","lastTransitionTime":"2025-11-29T00:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.604736 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.629523 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.654079 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.672140 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.695012 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.698665 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.698730 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.698749 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.698783 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.698842 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:22Z","lastTransitionTime":"2025-11-29T00:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.715680 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.739511 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.761722 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.785045 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.804370 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.804450 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.804490 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.804515 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.804530 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:22Z","lastTransitionTime":"2025-11-29T00:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.806265 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.825062 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.856187 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.874024 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.891646 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.908060 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.908126 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.908147 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.908178 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.908201 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:22Z","lastTransitionTime":"2025-11-29T00:00:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.913571 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.932081 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.942927 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.943110 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.943226 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:00:38.943187803 +0000 UTC m=+52.105081085 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.943313 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.943384 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.943464 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:38.943428359 +0000 UTC m=+52.105321661 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.943526 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.943514 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.943597 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:38.943581563 +0000 UTC m=+52.105474835 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.943728 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.943793 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.943797 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.943864 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.944107 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:38.944079866 +0000 UTC m=+52.105973148 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.943989 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.944214 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.944241 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:22 crc kubenswrapper[4931]: E1129 00:00:22.944345 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:38.944314752 +0000 UTC m=+52.106208024 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.960886 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:22 crc kubenswrapper[4931]: I1129 00:00:22.986437 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:22Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.005854 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.011028 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.011107 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.011130 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.011167 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.011192 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:23Z","lastTransitionTime":"2025-11-29T00:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.026116 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.045536 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.060986 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.083919 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.099348 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.112689 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.114113 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.114210 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.114240 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.114277 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.114305 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:23Z","lastTransitionTime":"2025-11-29T00:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.125584 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.145282 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.167094 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.212399 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:23 crc kubenswrapper[4931]: E1129 00:00:23.212588 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.213176 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:23 crc kubenswrapper[4931]: E1129 00:00:23.213366 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.220879 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.221018 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.221043 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.221071 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.221095 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:23Z","lastTransitionTime":"2025-11-29T00:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.323668 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.323707 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.323718 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.323736 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.323749 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:23Z","lastTransitionTime":"2025-11-29T00:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.426057 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.426108 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.426120 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.426142 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.426156 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:23Z","lastTransitionTime":"2025-11-29T00:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.528727 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.528793 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.528854 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.528882 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.528900 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:23Z","lastTransitionTime":"2025-11-29T00:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.632965 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.633034 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.633050 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.633077 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.633098 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:23Z","lastTransitionTime":"2025-11-29T00:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.735831 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.735886 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.735899 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.735923 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.735965 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:23Z","lastTransitionTime":"2025-11-29T00:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.839291 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.839359 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.839388 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.839413 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.839430 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:23Z","lastTransitionTime":"2025-11-29T00:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.942597 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.942785 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.942915 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.943069 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:23 crc kubenswrapper[4931]: I1129 00:00:23.943160 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:23Z","lastTransitionTime":"2025-11-29T00:00:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.048467 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.048544 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.048569 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.048601 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.048625 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:24Z","lastTransitionTime":"2025-11-29T00:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.152133 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.152587 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.152750 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.152934 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.153063 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:24Z","lastTransitionTime":"2025-11-29T00:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.212110 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:24 crc kubenswrapper[4931]: E1129 00:00:24.212357 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.256767 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.256879 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.256900 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.256926 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.256945 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:24Z","lastTransitionTime":"2025-11-29T00:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.359867 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.359958 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.359983 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.360024 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.360055 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:24Z","lastTransitionTime":"2025-11-29T00:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.463695 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.463781 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.463843 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.463885 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.463915 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:24Z","lastTransitionTime":"2025-11-29T00:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.532218 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/0.log" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.537279 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba" exitCode=1 Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.537355 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba"} Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.538613 4931 scope.go:117] "RemoveContainer" containerID="c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.565340 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.567413 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.567461 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.567479 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.567510 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.567531 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:24Z","lastTransitionTime":"2025-11-29T00:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.583345 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.607274 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.654276 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.671197 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.671273 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.671290 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.671314 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.671331 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:24Z","lastTransitionTime":"2025-11-29T00:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.685891 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.712168 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.732724 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.752233 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.768423 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.778603 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.778642 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.778651 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.778670 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.778680 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:24Z","lastTransitionTime":"2025-11-29T00:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.782663 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.800314 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.816588 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.835137 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:24Z\\\",\\\"message\\\":\\\"rnalversions/factory.go:140\\\\nI1129 00:00:24.030002 6267 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1129 00:00:24.030415 6267 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1129 00:00:24.030446 6267 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:24.030484 6267 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:24.030500 6267 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:24.030586 6267 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030657 6267 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030789 6267 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030986 6267 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.031344 6267 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:24.031396 6267 factory.go:656] Stopping watch factory\\\\nI1129 00:00:24.031410 6267 ovnkube.go:599] Stopped ovnkube\\\\nI1129 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.847133 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.857224 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:24Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.882666 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.882729 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.882746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.882773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.882796 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:24Z","lastTransitionTime":"2025-11-29T00:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.986479 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.986560 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.986585 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.986615 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:24 crc kubenswrapper[4931]: I1129 00:00:24.986653 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:24Z","lastTransitionTime":"2025-11-29T00:00:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.090604 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.090709 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.090728 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.090761 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.090786 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:25Z","lastTransitionTime":"2025-11-29T00:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.194788 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.194898 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.194927 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.194966 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.194994 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:25Z","lastTransitionTime":"2025-11-29T00:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.202570 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.212142 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.212320 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:25 crc kubenswrapper[4931]: E1129 00:00:25.212574 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:25 crc kubenswrapper[4931]: E1129 00:00:25.212927 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.227206 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.247605 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.271593 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.289506 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.300680 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.300736 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.300747 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.300768 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.300780 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:25Z","lastTransitionTime":"2025-11-29T00:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.315601 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.336104 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.350148 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.372127 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.396124 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.403381 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.403430 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.403442 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.403461 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.403471 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:25Z","lastTransitionTime":"2025-11-29T00:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.421564 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:24Z\\\",\\\"message\\\":\\\"rnalversions/factory.go:140\\\\nI1129 00:00:24.030002 6267 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1129 00:00:24.030415 6267 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1129 00:00:24.030446 6267 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:24.030484 6267 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:24.030500 6267 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:24.030586 6267 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030657 6267 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030789 6267 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030986 6267 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.031344 6267 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:24.031396 6267 factory.go:656] Stopping watch factory\\\\nI1129 00:00:24.031410 6267 ovnkube.go:599] Stopped ovnkube\\\\nI1129 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.444339 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.456390 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.467722 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.478391 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.489268 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.505422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.505472 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.505484 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.505510 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.505527 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:25Z","lastTransitionTime":"2025-11-29T00:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.545267 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/0.log" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.549503 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1"} Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.550199 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.581614 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.595653 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.609170 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.609230 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.609243 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.609265 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.609279 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:25Z","lastTransitionTime":"2025-11-29T00:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.615032 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.626784 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.639771 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.656652 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.674482 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.700040 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:24Z\\\",\\\"message\\\":\\\"rnalversions/factory.go:140\\\\nI1129 00:00:24.030002 6267 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1129 00:00:24.030415 6267 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1129 00:00:24.030446 6267 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:24.030484 6267 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:24.030500 6267 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:24.030586 6267 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030657 6267 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030789 6267 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030986 6267 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.031344 6267 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:24.031396 6267 factory.go:656] Stopping watch factory\\\\nI1129 00:00:24.031410 6267 ovnkube.go:599] Stopped ovnkube\\\\nI1129 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.712467 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.712522 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.712534 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.712554 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.712568 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:25Z","lastTransitionTime":"2025-11-29T00:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.716176 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.728250 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.741901 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.752451 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.768008 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.778162 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.791256 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:25Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.815851 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.815928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.815946 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.815976 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.815996 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:25Z","lastTransitionTime":"2025-11-29T00:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.919348 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.919403 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.919419 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.919446 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:25 crc kubenswrapper[4931]: I1129 00:00:25.919466 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:25Z","lastTransitionTime":"2025-11-29T00:00:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.022904 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.022996 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.023014 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.023044 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.023065 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:26Z","lastTransitionTime":"2025-11-29T00:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.126379 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.126454 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.126472 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.126499 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.126522 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:26Z","lastTransitionTime":"2025-11-29T00:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.211783 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:26 crc kubenswrapper[4931]: E1129 00:00:26.212041 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.230382 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.230465 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.230487 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.230515 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.230538 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:26Z","lastTransitionTime":"2025-11-29T00:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.333998 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.334082 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.334103 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.334139 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.334159 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:26Z","lastTransitionTime":"2025-11-29T00:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.437462 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.437561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.437586 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.437621 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.437643 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:26Z","lastTransitionTime":"2025-11-29T00:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.543059 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.543518 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.543774 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.544066 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.544231 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:26Z","lastTransitionTime":"2025-11-29T00:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.559318 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/1.log" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.561446 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/0.log" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.569176 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1" exitCode=1 Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.569268 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1"} Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.569440 4931 scope.go:117] "RemoveContainer" containerID="c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.570710 4931 scope.go:117] "RemoveContainer" containerID="41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1" Nov 29 00:00:26 crc kubenswrapper[4931]: E1129 00:00:26.571087 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.597659 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.615479 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.637128 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.648712 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.649080 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.649139 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.649172 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.649193 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:26Z","lastTransitionTime":"2025-11-29T00:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.655108 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.672582 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.694502 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.707863 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.734377 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:24Z\\\",\\\"message\\\":\\\"rnalversions/factory.go:140\\\\nI1129 00:00:24.030002 6267 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1129 00:00:24.030415 6267 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1129 00:00:24.030446 6267 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:24.030484 6267 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:24.030500 6267 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:24.030586 6267 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030657 6267 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030789 6267 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030986 6267 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.031344 6267 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:24.031396 6267 factory.go:656] Stopping watch factory\\\\nI1129 00:00:24.031410 6267 ovnkube.go:599] Stopped ovnkube\\\\nI1129 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:26Z\\\",\\\"message\\\":\\\"}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 00:00:25.905137 6393 services_controller.go:434] Service openshift-machine-api/machine-api-controllers retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-api-controllers openshift-machine-api 1cbb1d8a-02ea-4ab8-8f79-4dee9d158847 6869 0 2025-02-23 05:27:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:controller] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:machine-api-controllers-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077de97b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},Service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.753682 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.753756 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.753773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.753838 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.753863 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:26Z","lastTransitionTime":"2025-11-29T00:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.770255 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.790681 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.812948 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.834217 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.853219 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.857749 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.857854 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.858033 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.858060 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.858118 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:26Z","lastTransitionTime":"2025-11-29T00:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.872319 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.887871 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:26Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.961895 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.962020 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.962041 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.962068 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:26 crc kubenswrapper[4931]: I1129 00:00:26.962089 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:26Z","lastTransitionTime":"2025-11-29T00:00:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.066229 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.066275 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.066295 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.066319 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.066336 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:27Z","lastTransitionTime":"2025-11-29T00:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.170312 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.170377 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.170395 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.170422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.170443 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:27Z","lastTransitionTime":"2025-11-29T00:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.211347 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.211444 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:27 crc kubenswrapper[4931]: E1129 00:00:27.211510 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:27 crc kubenswrapper[4931]: E1129 00:00:27.211595 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.232675 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.256025 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.274713 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.274858 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.274882 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.274951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.274974 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:27Z","lastTransitionTime":"2025-11-29T00:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.281135 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.298780 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.322913 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.335185 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt"] Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.336228 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.339933 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.340170 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.344202 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.362436 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.378237 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.378262 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.378271 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.378288 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.378300 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:27Z","lastTransitionTime":"2025-11-29T00:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.383565 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.398561 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5767ea5-e04c-458d-a4bf-ca8941144fcb-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bs5dt\" (UID: \"d5767ea5-e04c-458d-a4bf-ca8941144fcb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.398659 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d5767ea5-e04c-458d-a4bf-ca8941144fcb-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bs5dt\" (UID: \"d5767ea5-e04c-458d-a4bf-ca8941144fcb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.398691 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bvsm\" (UniqueName: \"kubernetes.io/projected/d5767ea5-e04c-458d-a4bf-ca8941144fcb-kube-api-access-6bvsm\") pod \"ovnkube-control-plane-749d76644c-bs5dt\" (UID: \"d5767ea5-e04c-458d-a4bf-ca8941144fcb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.398728 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d5767ea5-e04c-458d-a4bf-ca8941144fcb-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bs5dt\" (UID: \"d5767ea5-e04c-458d-a4bf-ca8941144fcb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.402633 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.426836 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:24Z\\\",\\\"message\\\":\\\"rnalversions/factory.go:140\\\\nI1129 00:00:24.030002 6267 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1129 00:00:24.030415 6267 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1129 00:00:24.030446 6267 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:24.030484 6267 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:24.030500 6267 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:24.030586 6267 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030657 6267 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030789 6267 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030986 6267 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.031344 6267 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:24.031396 6267 factory.go:656] Stopping watch factory\\\\nI1129 00:00:24.031410 6267 ovnkube.go:599] Stopped ovnkube\\\\nI1129 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:26Z\\\",\\\"message\\\":\\\"}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 00:00:25.905137 6393 services_controller.go:434] Service openshift-machine-api/machine-api-controllers retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-api-controllers openshift-machine-api 1cbb1d8a-02ea-4ab8-8f79-4dee9d158847 6869 0 2025-02-23 05:27:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:controller] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:machine-api-controllers-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077de97b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},Service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.448237 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.462249 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.478530 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.481782 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.481869 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.481887 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.481913 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.481934 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:27Z","lastTransitionTime":"2025-11-29T00:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.499636 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d5767ea5-e04c-458d-a4bf-ca8941144fcb-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bs5dt\" (UID: \"d5767ea5-e04c-458d-a4bf-ca8941144fcb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.499697 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bvsm\" (UniqueName: \"kubernetes.io/projected/d5767ea5-e04c-458d-a4bf-ca8941144fcb-kube-api-access-6bvsm\") pod \"ovnkube-control-plane-749d76644c-bs5dt\" (UID: \"d5767ea5-e04c-458d-a4bf-ca8941144fcb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.499752 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d5767ea5-e04c-458d-a4bf-ca8941144fcb-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bs5dt\" (UID: \"d5767ea5-e04c-458d-a4bf-ca8941144fcb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.499798 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5767ea5-e04c-458d-a4bf-ca8941144fcb-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bs5dt\" (UID: \"d5767ea5-e04c-458d-a4bf-ca8941144fcb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.500449 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.500789 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5767ea5-e04c-458d-a4bf-ca8941144fcb-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bs5dt\" (UID: \"d5767ea5-e04c-458d-a4bf-ca8941144fcb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.501193 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d5767ea5-e04c-458d-a4bf-ca8941144fcb-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bs5dt\" (UID: \"d5767ea5-e04c-458d-a4bf-ca8941144fcb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.515328 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d5767ea5-e04c-458d-a4bf-ca8941144fcb-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bs5dt\" (UID: \"d5767ea5-e04c-458d-a4bf-ca8941144fcb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.523231 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bvsm\" (UniqueName: \"kubernetes.io/projected/d5767ea5-e04c-458d-a4bf-ca8941144fcb-kube-api-access-6bvsm\") pod \"ovnkube-control-plane-749d76644c-bs5dt\" (UID: \"d5767ea5-e04c-458d-a4bf-ca8941144fcb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.524402 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.548038 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.562023 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.574959 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/1.log" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.579623 4931 scope.go:117] "RemoveContainer" containerID="41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1" Nov 29 00:00:27 crc kubenswrapper[4931]: E1129 00:00:27.579853 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.580485 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.584314 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.584388 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.584402 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.584422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.584434 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:27Z","lastTransitionTime":"2025-11-29T00:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.595637 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.609618 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.634603 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.652393 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.658500 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.678949 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c282cb4302fe99ca1e640f61d35c80d2b9d09cb650ee848047208be6206a80ba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:24Z\\\",\\\"message\\\":\\\"rnalversions/factory.go:140\\\\nI1129 00:00:24.030002 6267 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1129 00:00:24.030415 6267 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1129 00:00:24.030446 6267 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:24.030484 6267 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:24.030500 6267 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:24.030586 6267 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030657 6267 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030789 6267 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.030986 6267 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1129 00:00:24.031344 6267 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:24.031396 6267 factory.go:656] Stopping watch factory\\\\nI1129 00:00:24.031410 6267 ovnkube.go:599] Stopped ovnkube\\\\nI1129 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:26Z\\\",\\\"message\\\":\\\"}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 00:00:25.905137 6393 services_controller.go:434] Service openshift-machine-api/machine-api-controllers retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-api-controllers openshift-machine-api 1cbb1d8a-02ea-4ab8-8f79-4dee9d158847 6869 0 2025-02-23 05:27:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:controller] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:machine-api-controllers-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077de97b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},Service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.687245 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.687297 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.687315 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.687343 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.687363 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:27Z","lastTransitionTime":"2025-11-29T00:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.710987 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.730533 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.748134 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.760617 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.774053 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.788705 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.792477 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.792545 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.792567 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.792611 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.792632 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:27Z","lastTransitionTime":"2025-11-29T00:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.808020 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.823287 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.839209 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.853588 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.868350 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.881686 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.896321 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.896375 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.896389 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.896413 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.896427 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:27Z","lastTransitionTime":"2025-11-29T00:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.900626 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.918671 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.932805 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.944483 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.957531 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.971591 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.996043 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.999448 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.999524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.999548 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.999576 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:27 crc kubenswrapper[4931]: I1129 00:00:27.999597 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:27Z","lastTransitionTime":"2025-11-29T00:00:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.009276 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.027804 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.047495 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.065577 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:26Z\\\",\\\"message\\\":\\\"}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 00:00:25.905137 6393 services_controller.go:434] Service openshift-machine-api/machine-api-controllers retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-api-controllers openshift-machine-api 1cbb1d8a-02ea-4ab8-8f79-4dee9d158847 6869 0 2025-02-23 05:27:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:controller] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:machine-api-controllers-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077de97b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},Service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.093065 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.101916 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.102000 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.102025 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.102058 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.102085 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:28Z","lastTransitionTime":"2025-11-29T00:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.205637 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.205710 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.205729 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.205757 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.205779 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:28Z","lastTransitionTime":"2025-11-29T00:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.212012 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:28 crc kubenswrapper[4931]: E1129 00:00:28.212198 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.309367 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.309433 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.309450 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.309476 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.309496 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:28Z","lastTransitionTime":"2025-11-29T00:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.412543 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.412856 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.412894 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.412922 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.412940 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:28Z","lastTransitionTime":"2025-11-29T00:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.516356 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.516410 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.516422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.516443 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.516458 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:28Z","lastTransitionTime":"2025-11-29T00:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.584557 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" event={"ID":"d5767ea5-e04c-458d-a4bf-ca8941144fcb","Type":"ContainerStarted","Data":"2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32"} Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.584607 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" event={"ID":"d5767ea5-e04c-458d-a4bf-ca8941144fcb","Type":"ContainerStarted","Data":"6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad"} Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.584617 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" event={"ID":"d5767ea5-e04c-458d-a4bf-ca8941144fcb","Type":"ContainerStarted","Data":"90f773572644fbff1968461efbe22580c8fc478b0880f0e736f502f8780b45b5"} Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.605676 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.619330 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.619390 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.619409 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.619440 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.619460 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:28Z","lastTransitionTime":"2025-11-29T00:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.620026 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.650424 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.670297 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.682956 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.694923 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.721075 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:26Z\\\",\\\"message\\\":\\\"}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 00:00:25.905137 6393 services_controller.go:434] Service openshift-machine-api/machine-api-controllers retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-api-controllers openshift-machine-api 1cbb1d8a-02ea-4ab8-8f79-4dee9d158847 6869 0 2025-02-23 05:27:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:controller] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:machine-api-controllers-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077de97b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},Service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.721652 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.721701 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.721712 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.721732 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.721746 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:28Z","lastTransitionTime":"2025-11-29T00:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.765219 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.786391 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.807644 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.822446 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.823996 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.824023 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.824033 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.824051 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.824060 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:28Z","lastTransitionTime":"2025-11-29T00:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.838523 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.852352 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.867016 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.880369 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.891472 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:28Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.926667 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.926710 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.926718 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.926735 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:28 crc kubenswrapper[4931]: I1129 00:00:28.926748 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:28Z","lastTransitionTime":"2025-11-29T00:00:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.029366 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.029417 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.029430 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.029455 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.029468 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:29Z","lastTransitionTime":"2025-11-29T00:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.132975 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.133300 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.133402 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.133495 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.133574 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:29Z","lastTransitionTime":"2025-11-29T00:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.206635 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-7dwtb"] Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.207536 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:29 crc kubenswrapper[4931]: E1129 00:00:29.207680 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.211279 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.211284 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:29 crc kubenswrapper[4931]: E1129 00:00:29.211490 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:29 crc kubenswrapper[4931]: E1129 00:00:29.211623 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.226617 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.236321 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.236409 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.236430 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.236454 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.236471 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:29Z","lastTransitionTime":"2025-11-29T00:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.243850 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.265447 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.287946 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.304529 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.326619 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.326703 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-848rn\" (UniqueName: \"kubernetes.io/projected/5e38ad69-f5fb-415d-ad76-e899aed032a6-kube-api-access-848rn\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.327945 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.339701 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.340052 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.340230 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.340459 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.340614 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:29Z","lastTransitionTime":"2025-11-29T00:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.347618 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.363542 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.380192 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.411042 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:26Z\\\",\\\"message\\\":\\\"}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 00:00:25.905137 6393 services_controller.go:434] Service openshift-machine-api/machine-api-controllers retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-api-controllers openshift-machine-api 1cbb1d8a-02ea-4ab8-8f79-4dee9d158847 6869 0 2025-02-23 05:27:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:controller] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:machine-api-controllers-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077de97b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},Service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.428235 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.428352 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-848rn\" (UniqueName: \"kubernetes.io/projected/5e38ad69-f5fb-415d-ad76-e899aed032a6-kube-api-access-848rn\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:29 crc kubenswrapper[4931]: E1129 00:00:29.428518 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:00:29 crc kubenswrapper[4931]: E1129 00:00:29.428638 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs podName:5e38ad69-f5fb-415d-ad76-e899aed032a6 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:29.928601417 +0000 UTC m=+43.090494719 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs") pod "network-metrics-daemon-7dwtb" (UID: "5e38ad69-f5fb-415d-ad76-e899aed032a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.444191 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.444238 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.444288 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.444316 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.444340 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:29Z","lastTransitionTime":"2025-11-29T00:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.448195 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.465449 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-848rn\" (UniqueName: \"kubernetes.io/projected/5e38ad69-f5fb-415d-ad76-e899aed032a6-kube-api-access-848rn\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.473054 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.493017 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.509007 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.522792 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.546937 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.547741 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.547796 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.547838 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.547866 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.547917 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:29Z","lastTransitionTime":"2025-11-29T00:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.565110 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:29Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.650989 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.651044 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.651056 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.651075 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.651088 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:29Z","lastTransitionTime":"2025-11-29T00:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.754098 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.754169 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.754188 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.754213 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.754234 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:29Z","lastTransitionTime":"2025-11-29T00:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.857589 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.857646 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.857663 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.857691 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.857709 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:29Z","lastTransitionTime":"2025-11-29T00:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.934521 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:29 crc kubenswrapper[4931]: E1129 00:00:29.934770 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:00:29 crc kubenswrapper[4931]: E1129 00:00:29.934884 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs podName:5e38ad69-f5fb-415d-ad76-e899aed032a6 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:30.9348572 +0000 UTC m=+44.096750442 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs") pod "network-metrics-daemon-7dwtb" (UID: "5e38ad69-f5fb-415d-ad76-e899aed032a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.960032 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.960111 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.960177 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.960200 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:29 crc kubenswrapper[4931]: I1129 00:00:29.960211 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:29Z","lastTransitionTime":"2025-11-29T00:00:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.063014 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.063097 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.063118 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.063146 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.063168 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:30Z","lastTransitionTime":"2025-11-29T00:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.166335 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.166397 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.166416 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.166444 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.166463 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:30Z","lastTransitionTime":"2025-11-29T00:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.212195 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:30 crc kubenswrapper[4931]: E1129 00:00:30.212420 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.270255 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.270354 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.270378 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.270418 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.270447 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:30Z","lastTransitionTime":"2025-11-29T00:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.375987 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.376078 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.376107 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.376149 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.376176 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:30Z","lastTransitionTime":"2025-11-29T00:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.479887 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.479931 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.479940 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.479957 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.479966 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:30Z","lastTransitionTime":"2025-11-29T00:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.583492 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.583573 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.583595 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.583624 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.583644 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:30Z","lastTransitionTime":"2025-11-29T00:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.687133 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.687202 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.687220 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.687249 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.687269 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:30Z","lastTransitionTime":"2025-11-29T00:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.790044 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.790096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.790114 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.790136 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.790152 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:30Z","lastTransitionTime":"2025-11-29T00:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.893738 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.893783 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.893800 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.893846 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.893863 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:30Z","lastTransitionTime":"2025-11-29T00:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.948573 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:30 crc kubenswrapper[4931]: E1129 00:00:30.948964 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:00:30 crc kubenswrapper[4931]: E1129 00:00:30.949132 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs podName:5e38ad69-f5fb-415d-ad76-e899aed032a6 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:32.949091841 +0000 UTC m=+46.110985103 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs") pod "network-metrics-daemon-7dwtb" (UID: "5e38ad69-f5fb-415d-ad76-e899aed032a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.996776 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.996856 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.996879 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.996906 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:30 crc kubenswrapper[4931]: I1129 00:00:30.996923 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:30Z","lastTransitionTime":"2025-11-29T00:00:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.100563 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.100657 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.100683 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.100718 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.100745 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.194716 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.194798 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.194851 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.194899 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.194921 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.211695 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:31 crc kubenswrapper[4931]: E1129 00:00:31.212002 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.212547 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:31 crc kubenswrapper[4931]: E1129 00:00:31.212659 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.212878 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:31 crc kubenswrapper[4931]: E1129 00:00:31.213001 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:31 crc kubenswrapper[4931]: E1129 00:00:31.216869 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:31Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.223958 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.224012 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.224030 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.224056 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.224079 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: E1129 00:00:31.246373 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:31Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.251517 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.251548 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.251562 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.251582 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.251596 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: E1129 00:00:31.271355 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:31Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.281920 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.281971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.281988 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.282008 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.282027 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: E1129 00:00:31.299965 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:31Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.305430 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.305478 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.305497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.305525 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.305545 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: E1129 00:00:31.327079 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:31Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:31 crc kubenswrapper[4931]: E1129 00:00:31.327390 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.329650 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.329725 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.329753 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.329787 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.329844 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.432759 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.432892 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.432918 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.432948 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.432973 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.536245 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.536293 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.536305 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.536323 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.536336 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.640196 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.640265 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.640295 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.640327 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.640350 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.744253 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.744649 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.744790 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.744972 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.745115 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.848894 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.848978 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.848999 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.849031 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.849052 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.951951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.952030 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.952049 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.952076 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:31 crc kubenswrapper[4931]: I1129 00:00:31.952098 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:31Z","lastTransitionTime":"2025-11-29T00:00:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.055510 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.055593 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.055616 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.055653 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.055676 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:32Z","lastTransitionTime":"2025-11-29T00:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.158528 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.158622 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.158647 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.158682 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.158734 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:32Z","lastTransitionTime":"2025-11-29T00:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.211471 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:32 crc kubenswrapper[4931]: E1129 00:00:32.211675 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.262254 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.262675 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.262799 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.262950 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.263036 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:32Z","lastTransitionTime":"2025-11-29T00:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.365847 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.365881 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.365892 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.365909 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.365924 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:32Z","lastTransitionTime":"2025-11-29T00:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.469695 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.469751 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.469768 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.469792 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.469824 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:32Z","lastTransitionTime":"2025-11-29T00:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.574004 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.574091 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.574110 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.574140 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.574162 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:32Z","lastTransitionTime":"2025-11-29T00:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.677361 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.677442 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.677465 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.677493 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.677516 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:32Z","lastTransitionTime":"2025-11-29T00:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.781051 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.781537 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.781704 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.781908 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.782114 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:32Z","lastTransitionTime":"2025-11-29T00:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.886322 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.886390 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.886409 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.886484 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.886548 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:32Z","lastTransitionTime":"2025-11-29T00:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.975476 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:32 crc kubenswrapper[4931]: E1129 00:00:32.975839 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:00:32 crc kubenswrapper[4931]: E1129 00:00:32.975961 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs podName:5e38ad69-f5fb-415d-ad76-e899aed032a6 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:36.975929989 +0000 UTC m=+50.137823251 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs") pod "network-metrics-daemon-7dwtb" (UID: "5e38ad69-f5fb-415d-ad76-e899aed032a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.990451 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.990508 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.990518 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.990538 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:32 crc kubenswrapper[4931]: I1129 00:00:32.990549 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:32Z","lastTransitionTime":"2025-11-29T00:00:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.093005 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.093064 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.093078 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.093100 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.093115 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:33Z","lastTransitionTime":"2025-11-29T00:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.195698 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.195743 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.195754 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.195773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.195785 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:33Z","lastTransitionTime":"2025-11-29T00:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.211292 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.211427 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:33 crc kubenswrapper[4931]: E1129 00:00:33.211540 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.211596 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:33 crc kubenswrapper[4931]: E1129 00:00:33.211774 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:33 crc kubenswrapper[4931]: E1129 00:00:33.211922 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.299213 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.299268 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.299285 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.299305 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.299321 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:33Z","lastTransitionTime":"2025-11-29T00:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.402437 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.402491 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.402506 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.402528 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.402543 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:33Z","lastTransitionTime":"2025-11-29T00:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.505977 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.506405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.506515 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.506633 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.506742 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:33Z","lastTransitionTime":"2025-11-29T00:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.609735 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.610180 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.610319 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.610422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.610503 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:33Z","lastTransitionTime":"2025-11-29T00:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.714048 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.714502 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.714611 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.714727 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.714831 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:33Z","lastTransitionTime":"2025-11-29T00:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.818215 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.818603 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.818750 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.819087 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.819221 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:33Z","lastTransitionTime":"2025-11-29T00:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.922730 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.923357 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.923381 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.923406 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:33 crc kubenswrapper[4931]: I1129 00:00:33.923423 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:33Z","lastTransitionTime":"2025-11-29T00:00:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.027711 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.028398 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.028551 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.028739 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.028937 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:34Z","lastTransitionTime":"2025-11-29T00:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.132592 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.132637 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.132650 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.132667 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.132678 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:34Z","lastTransitionTime":"2025-11-29T00:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.212051 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:34 crc kubenswrapper[4931]: E1129 00:00:34.212336 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.236167 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.236426 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.236456 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.236493 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.236518 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:34Z","lastTransitionTime":"2025-11-29T00:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.340446 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.340481 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.340490 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.340506 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.340518 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:34Z","lastTransitionTime":"2025-11-29T00:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.443355 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.443402 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.443414 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.443431 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.443443 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:34Z","lastTransitionTime":"2025-11-29T00:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.546796 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.546865 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.546877 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.546895 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.546911 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:34Z","lastTransitionTime":"2025-11-29T00:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.651111 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.651800 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.651941 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.652067 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.652164 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:34Z","lastTransitionTime":"2025-11-29T00:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.755615 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.756097 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.756282 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.756456 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.756633 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:34Z","lastTransitionTime":"2025-11-29T00:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.860602 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.861126 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.861273 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.861400 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.861513 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:34Z","lastTransitionTime":"2025-11-29T00:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.965352 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.965422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.965441 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.965467 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:34 crc kubenswrapper[4931]: I1129 00:00:34.965490 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:34Z","lastTransitionTime":"2025-11-29T00:00:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.068866 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.068931 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.068988 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.069017 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.069032 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:35Z","lastTransitionTime":"2025-11-29T00:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.172009 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.172080 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.172097 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.172124 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.172147 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:35Z","lastTransitionTime":"2025-11-29T00:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.212334 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.212409 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.212333 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:35 crc kubenswrapper[4931]: E1129 00:00:35.212515 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:35 crc kubenswrapper[4931]: E1129 00:00:35.212589 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:35 crc kubenswrapper[4931]: E1129 00:00:35.212745 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.275331 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.275413 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.275437 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.275468 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.275495 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:35Z","lastTransitionTime":"2025-11-29T00:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.408394 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.408456 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.408477 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.408506 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.408537 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:35Z","lastTransitionTime":"2025-11-29T00:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.511094 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.511162 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.511182 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.511210 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.511228 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:35Z","lastTransitionTime":"2025-11-29T00:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.612945 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.612991 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.613003 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.613020 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.613030 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:35Z","lastTransitionTime":"2025-11-29T00:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.716588 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.717037 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.717244 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.717593 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.717750 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:35Z","lastTransitionTime":"2025-11-29T00:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.822692 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.822745 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.822755 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.822774 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.822783 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:35Z","lastTransitionTime":"2025-11-29T00:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.926704 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.926767 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.926786 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.926848 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:35 crc kubenswrapper[4931]: I1129 00:00:35.926867 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:35Z","lastTransitionTime":"2025-11-29T00:00:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.030519 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.030590 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.030610 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.030638 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.030659 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:36Z","lastTransitionTime":"2025-11-29T00:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.132987 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.133390 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.133551 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.133719 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.133878 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:36Z","lastTransitionTime":"2025-11-29T00:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.211616 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:36 crc kubenswrapper[4931]: E1129 00:00:36.211897 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.237669 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.237768 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.237786 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.237882 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.237901 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:36Z","lastTransitionTime":"2025-11-29T00:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.341021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.341194 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.341247 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.341282 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.341305 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:36Z","lastTransitionTime":"2025-11-29T00:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.444561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.445043 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.445269 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.445492 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.445721 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:36Z","lastTransitionTime":"2025-11-29T00:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.549652 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.549730 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.549753 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.549779 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.549797 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:36Z","lastTransitionTime":"2025-11-29T00:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.653963 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.654052 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.654072 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.654101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.654121 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:36Z","lastTransitionTime":"2025-11-29T00:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.757214 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.757290 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.757308 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.757339 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.757394 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:36Z","lastTransitionTime":"2025-11-29T00:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.860658 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.860783 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.860847 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.860890 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.860924 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:36Z","lastTransitionTime":"2025-11-29T00:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.964256 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.964325 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.964349 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.964380 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:36 crc kubenswrapper[4931]: I1129 00:00:36.964403 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:36Z","lastTransitionTime":"2025-11-29T00:00:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.026652 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:37 crc kubenswrapper[4931]: E1129 00:00:37.026896 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:00:37 crc kubenswrapper[4931]: E1129 00:00:37.026988 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs podName:5e38ad69-f5fb-415d-ad76-e899aed032a6 nodeName:}" failed. No retries permitted until 2025-11-29 00:00:45.026957678 +0000 UTC m=+58.188850940 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs") pod "network-metrics-daemon-7dwtb" (UID: "5e38ad69-f5fb-415d-ad76-e899aed032a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.068340 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.069139 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.069190 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.069229 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.069255 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:37Z","lastTransitionTime":"2025-11-29T00:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.172673 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.172750 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.172775 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.172845 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.172868 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:37Z","lastTransitionTime":"2025-11-29T00:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.212165 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.212235 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.212257 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:37 crc kubenswrapper[4931]: E1129 00:00:37.212340 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:37 crc kubenswrapper[4931]: E1129 00:00:37.212480 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:37 crc kubenswrapper[4931]: E1129 00:00:37.212652 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.235079 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.254514 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.275371 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.275407 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.275422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.275442 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.275456 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:37Z","lastTransitionTime":"2025-11-29T00:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.278277 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.300833 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.316795 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.338996 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.359433 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.375386 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.378332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.378436 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.378457 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.378486 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.378505 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:37Z","lastTransitionTime":"2025-11-29T00:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.403781 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.427005 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.462497 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:26Z\\\",\\\"message\\\":\\\"}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 00:00:25.905137 6393 services_controller.go:434] Service openshift-machine-api/machine-api-controllers retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-api-controllers openshift-machine-api 1cbb1d8a-02ea-4ab8-8f79-4dee9d158847 6869 0 2025-02-23 05:27:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:controller] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:machine-api-controllers-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077de97b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},Service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.481381 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.481438 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.481457 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.481482 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.481502 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:37Z","lastTransitionTime":"2025-11-29T00:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.504467 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.524364 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.545234 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.566541 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.584271 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.584322 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.584341 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.584369 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.584425 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:37Z","lastTransitionTime":"2025-11-29T00:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.592162 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.608997 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:37Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.687615 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.687655 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.687667 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.687683 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.687693 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:37Z","lastTransitionTime":"2025-11-29T00:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.793369 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.793411 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.793422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.793438 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.793449 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:37Z","lastTransitionTime":"2025-11-29T00:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.896418 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.896461 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.896478 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.896498 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:37 crc kubenswrapper[4931]: I1129 00:00:37.896513 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:37Z","lastTransitionTime":"2025-11-29T00:00:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.001298 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.001409 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.001435 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.001465 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.001486 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:38Z","lastTransitionTime":"2025-11-29T00:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.104802 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.104893 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.104911 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.104938 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.104956 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:38Z","lastTransitionTime":"2025-11-29T00:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.208210 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.208277 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.208302 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.208332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.208356 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:38Z","lastTransitionTime":"2025-11-29T00:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.211878 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.212068 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.312083 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.312150 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.312170 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.312195 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.312214 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:38Z","lastTransitionTime":"2025-11-29T00:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.414707 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.414761 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.414772 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.414794 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.414824 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:38Z","lastTransitionTime":"2025-11-29T00:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.517738 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.518005 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.518045 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.518075 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.518095 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:38Z","lastTransitionTime":"2025-11-29T00:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.621384 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.621430 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.621441 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.621459 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.621472 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:38Z","lastTransitionTime":"2025-11-29T00:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.725227 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.725704 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.725863 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.726002 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.726116 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:38Z","lastTransitionTime":"2025-11-29T00:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.829713 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.829797 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.829850 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.829878 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.829899 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:38Z","lastTransitionTime":"2025-11-29T00:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.933219 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.933620 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.933856 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.934099 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.934249 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:38Z","lastTransitionTime":"2025-11-29T00:00:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.951954 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952190 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:01:10.952156224 +0000 UTC m=+84.114049476 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.952310 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.952380 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.952421 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:38 crc kubenswrapper[4931]: I1129 00:00:38.952459 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952520 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952580 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952589 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:01:10.952575835 +0000 UTC m=+84.114469077 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952668 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952690 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:01:10.952669077 +0000 UTC m=+84.114562349 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952695 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952725 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952750 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952844 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952850 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 00:01:10.9527861 +0000 UTC m=+84.114679372 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952873 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:38 crc kubenswrapper[4931]: E1129 00:00:38.952988 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 00:01:10.952954185 +0000 UTC m=+84.114847457 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.038211 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.038299 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.038332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.038428 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.038462 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:39Z","lastTransitionTime":"2025-11-29T00:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.141417 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.141475 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.141489 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.141511 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.141526 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:39Z","lastTransitionTime":"2025-11-29T00:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.212416 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.212512 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.213015 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:39 crc kubenswrapper[4931]: E1129 00:00:39.213180 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:39 crc kubenswrapper[4931]: E1129 00:00:39.213420 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:39 crc kubenswrapper[4931]: E1129 00:00:39.213608 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.213661 4931 scope.go:117] "RemoveContainer" containerID="41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.246885 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.247757 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.247911 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.248048 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.248190 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:39Z","lastTransitionTime":"2025-11-29T00:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.351665 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.352244 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.352387 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.352456 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.352516 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:39Z","lastTransitionTime":"2025-11-29T00:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.455586 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.455637 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.455646 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.455663 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.455673 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:39Z","lastTransitionTime":"2025-11-29T00:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.558837 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.559502 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.559518 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.559537 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.559552 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:39Z","lastTransitionTime":"2025-11-29T00:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.634934 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/1.log" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.639916 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232"} Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.640730 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.661476 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.661513 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.661522 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.661559 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.661569 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:39Z","lastTransitionTime":"2025-11-29T00:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.665383 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.682902 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.700792 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.730832 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:26Z\\\",\\\"message\\\":\\\"}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 00:00:25.905137 6393 services_controller.go:434] Service openshift-machine-api/machine-api-controllers retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-api-controllers openshift-machine-api 1cbb1d8a-02ea-4ab8-8f79-4dee9d158847 6869 0 2025-02-23 05:27:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:controller] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:machine-api-controllers-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077de97b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},Service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.755319 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.764024 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.764086 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.764102 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.764127 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.764142 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:39Z","lastTransitionTime":"2025-11-29T00:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.769525 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.782001 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.796776 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.809384 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.825896 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.840145 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.851610 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.867536 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.867605 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.867619 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.867641 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.867675 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:39Z","lastTransitionTime":"2025-11-29T00:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.869247 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.885596 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.900444 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.917273 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.928865 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:39Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.970081 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.970137 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.970152 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.970174 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:39 crc kubenswrapper[4931]: I1129 00:00:39.970190 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:39Z","lastTransitionTime":"2025-11-29T00:00:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.074037 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.074106 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.074129 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.074167 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.074193 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:40Z","lastTransitionTime":"2025-11-29T00:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.176589 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.176640 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.176653 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.176675 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.176690 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:40Z","lastTransitionTime":"2025-11-29T00:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.212007 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:40 crc kubenswrapper[4931]: E1129 00:00:40.212187 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.281076 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.281182 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.281201 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.281228 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.281247 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:40Z","lastTransitionTime":"2025-11-29T00:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.385192 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.385272 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.385300 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.385332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.385355 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:40Z","lastTransitionTime":"2025-11-29T00:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.489307 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.489423 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.489449 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.489482 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.489508 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:40Z","lastTransitionTime":"2025-11-29T00:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.592657 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.592716 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.592736 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.592762 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.592781 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:40Z","lastTransitionTime":"2025-11-29T00:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.646464 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/2.log" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.648014 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/1.log" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.652106 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232" exitCode=1 Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.652218 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232"} Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.652450 4931 scope.go:117] "RemoveContainer" containerID="41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.653229 4931 scope.go:117] "RemoveContainer" containerID="32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232" Nov 29 00:00:40 crc kubenswrapper[4931]: E1129 00:00:40.653499 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.679965 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.696457 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.696561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.696586 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.696677 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.696736 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:40Z","lastTransitionTime":"2025-11-29T00:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.703434 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.722307 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.746022 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.764052 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.788280 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.801643 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.801692 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.801710 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.801738 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.801759 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:40Z","lastTransitionTime":"2025-11-29T00:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.811100 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.829603 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.851143 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.885050 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41ac91e8dc682bf6050c4e3400ba73eaaaff5ad91d534a2e13223d17d1d0f5f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:26Z\\\",\\\"message\\\":\\\"}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1129 00:00:25.905137 6393 services_controller.go:434] Service openshift-machine-api/machine-api-controllers retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-api-controllers openshift-machine-api 1cbb1d8a-02ea-4ab8-8f79-4dee9d158847 6869 0 2025-02-23 05:27:48 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:controller] map[capability.openshift.io/name:MachineAPI exclude.release.openshift.io/internal-openshift-hosted:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-secret-name:machine-api-controllers-tls service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0077de97b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:machine-mtrc,Protocol:TCP,Port:8441,TargetPort:{1 0 machine-mtrc},NodePort:0,AppProtocol:nil,},ServicePort{Name:machineset-mtrc,Protocol:TCP,Port:8442,TargetPort:{1 0 machineset-mtrc},NodePort:0,AppProtocol:nil,},Service\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:40Z\\\",\\\"message\\\":\\\" 6599 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:40.145441 6599 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:00:40.145452 6599 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:40.145480 6599 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 00:00:40.145520 6599 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:40.145531 6599 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 00:00:40.145539 6599 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 00:00:40.145571 6599 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:40.145630 6599 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:00:40.145668 6599 factory.go:656] Stopping watch factory\\\\nI1129 00:00:40.145692 6599 ovnkube.go:599] Stopped ovnkube\\\\nI1129 00:00:40.145728 6599 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:00:40.145738 6599 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 00:00:40.145747 6599 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 00:00:40.145792 6599 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF1129 00:00:40.145877 6599 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.905082 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.905152 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.905169 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.905195 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.905214 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:40Z","lastTransitionTime":"2025-11-29T00:00:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.923885 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.955350 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.975668 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:40 crc kubenswrapper[4931]: I1129 00:00:40.996498 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:40Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.007981 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.008044 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.008062 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.008089 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.008107 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.017265 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.044286 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.064318 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.111391 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.111467 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.111494 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.111529 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.111552 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.212247 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.212313 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:41 crc kubenswrapper[4931]: E1129 00:00:41.212489 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.212548 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:41 crc kubenswrapper[4931]: E1129 00:00:41.212633 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:41 crc kubenswrapper[4931]: E1129 00:00:41.212772 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.215920 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.216007 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.216040 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.216067 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.216085 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.319426 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.319504 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.319530 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.319565 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.319594 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.423187 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.423257 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.423328 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.423363 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.423390 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.526837 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.526929 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.526950 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.526981 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.527004 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.634251 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.634326 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.634344 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.634370 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.634389 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.636035 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.636175 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.636289 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.636376 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.636480 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: E1129 00:00:41.655564 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.659166 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/2.log" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.660687 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.660727 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.660744 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.660769 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.660787 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.664318 4931 scope.go:117] "RemoveContainer" containerID="32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232" Nov 29 00:00:41 crc kubenswrapper[4931]: E1129 00:00:41.664643 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" Nov 29 00:00:41 crc kubenswrapper[4931]: E1129 00:00:41.683117 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.685991 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.687788 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.687880 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.687904 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.687942 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.687962 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.703133 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: E1129 00:00:41.705753 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.710607 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.710736 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.710762 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.710791 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.710835 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.719363 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: E1129 00:00:41.727470 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.731051 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.731097 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.731110 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.731134 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.731148 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.743306 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: E1129 00:00:41.748300 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: E1129 00:00:41.748408 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.750845 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.750884 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.750897 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.750915 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.750927 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.761132 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.776628 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.794138 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.808798 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.828765 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.846198 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.853727 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.853861 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.853932 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.853969 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.853992 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.868470 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:40Z\\\",\\\"message\\\":\\\" 6599 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:40.145441 6599 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:00:40.145452 6599 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:40.145480 6599 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 00:00:40.145520 6599 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:40.145531 6599 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 00:00:40.145539 6599 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 00:00:40.145571 6599 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:40.145630 6599 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:00:40.145668 6599 factory.go:656] Stopping watch factory\\\\nI1129 00:00:40.145692 6599 ovnkube.go:599] Stopped ovnkube\\\\nI1129 00:00:40.145728 6599 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:00:40.145738 6599 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 00:00:40.145747 6599 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 00:00:40.145792 6599 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF1129 00:00:40.145877 6599 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.883997 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.902149 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.916933 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.931537 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.952532 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.957299 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.957515 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.957718 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.957932 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.958135 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:41Z","lastTransitionTime":"2025-11-29T00:00:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:41 crc kubenswrapper[4931]: I1129 00:00:41.968705 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:41Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.061630 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.061712 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.061739 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.061775 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.061801 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:42Z","lastTransitionTime":"2025-11-29T00:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.165803 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.166290 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.166461 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.166620 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.166760 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:42Z","lastTransitionTime":"2025-11-29T00:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.211989 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:42 crc kubenswrapper[4931]: E1129 00:00:42.212184 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.270599 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.270669 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.270686 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.270714 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.270733 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:42Z","lastTransitionTime":"2025-11-29T00:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.374104 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.374210 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.374237 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.374270 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.374296 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:42Z","lastTransitionTime":"2025-11-29T00:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.477780 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.477872 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.477897 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.477929 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.477954 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:42Z","lastTransitionTime":"2025-11-29T00:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.581716 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.581785 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.581842 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.581878 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.581899 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:42Z","lastTransitionTime":"2025-11-29T00:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.685625 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.685700 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.685721 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.685751 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.685771 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:42Z","lastTransitionTime":"2025-11-29T00:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.789169 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.789243 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.789262 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.789293 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.789313 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:42Z","lastTransitionTime":"2025-11-29T00:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.893152 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.893192 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.893201 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.893222 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.893234 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:42Z","lastTransitionTime":"2025-11-29T00:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.997399 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.997488 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.997508 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.997537 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:42 crc kubenswrapper[4931]: I1129 00:00:42.997555 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:42Z","lastTransitionTime":"2025-11-29T00:00:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.007795 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.023195 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.030865 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.047367 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.069638 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.091385 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.107085 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.107175 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.107199 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.107234 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.107257 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:43Z","lastTransitionTime":"2025-11-29T00:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.113006 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.152177 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:40Z\\\",\\\"message\\\":\\\" 6599 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:40.145441 6599 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:00:40.145452 6599 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:40.145480 6599 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 00:00:40.145520 6599 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:40.145531 6599 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 00:00:40.145539 6599 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 00:00:40.145571 6599 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:40.145630 6599 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:00:40.145668 6599 factory.go:656] Stopping watch factory\\\\nI1129 00:00:40.145692 6599 ovnkube.go:599] Stopped ovnkube\\\\nI1129 00:00:40.145728 6599 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:00:40.145738 6599 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 00:00:40.145747 6599 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 00:00:40.145792 6599 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF1129 00:00:40.145877 6599 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.174559 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.187517 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.200174 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.210116 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.210418 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.210483 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.210569 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.210641 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:43Z","lastTransitionTime":"2025-11-29T00:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.211317 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:43 crc kubenswrapper[4931]: E1129 00:00:43.211514 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.211334 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.211330 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:43 crc kubenswrapper[4931]: E1129 00:00:43.211720 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:43 crc kubenswrapper[4931]: E1129 00:00:43.211912 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.215868 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.227575 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.240744 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.257801 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.272747 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.289259 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.304120 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.314103 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.314162 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.314179 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.314207 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.314228 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:43Z","lastTransitionTime":"2025-11-29T00:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.320783 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:43Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.417853 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.417948 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.417975 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.418014 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.418036 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:43Z","lastTransitionTime":"2025-11-29T00:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.521191 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.521677 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.521924 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.522096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.522243 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:43Z","lastTransitionTime":"2025-11-29T00:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.625892 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.626302 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.626448 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.626593 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.626726 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:43Z","lastTransitionTime":"2025-11-29T00:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.730764 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.730859 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.730880 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.730909 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.730930 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:43Z","lastTransitionTime":"2025-11-29T00:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.834703 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.834756 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.834774 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.835071 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.835095 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:43Z","lastTransitionTime":"2025-11-29T00:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.938613 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.938715 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.938739 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.938776 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:43 crc kubenswrapper[4931]: I1129 00:00:43.938840 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:43Z","lastTransitionTime":"2025-11-29T00:00:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.043770 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.043870 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.043910 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.043944 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.043966 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:44Z","lastTransitionTime":"2025-11-29T00:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.148599 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.148669 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.148687 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.148714 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.148732 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:44Z","lastTransitionTime":"2025-11-29T00:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.211676 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:44 crc kubenswrapper[4931]: E1129 00:00:44.211991 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.252854 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.252943 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.252967 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.253002 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.253027 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:44Z","lastTransitionTime":"2025-11-29T00:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.357421 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.357502 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.357530 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.357565 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.357594 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:44Z","lastTransitionTime":"2025-11-29T00:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.460534 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.460607 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.460627 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.460654 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.460675 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:44Z","lastTransitionTime":"2025-11-29T00:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.564287 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.564355 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.564373 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.564404 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.564424 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:44Z","lastTransitionTime":"2025-11-29T00:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.667098 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.667154 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.667167 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.667186 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.667199 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:44Z","lastTransitionTime":"2025-11-29T00:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.771100 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.771168 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.771182 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.771205 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.771221 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:44Z","lastTransitionTime":"2025-11-29T00:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.874418 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.874505 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.874532 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.874566 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.874592 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:44Z","lastTransitionTime":"2025-11-29T00:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.978468 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.978534 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.978551 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.978576 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:44 crc kubenswrapper[4931]: I1129 00:00:44.978604 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:44Z","lastTransitionTime":"2025-11-29T00:00:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.029223 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:45 crc kubenswrapper[4931]: E1129 00:00:45.029418 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:00:45 crc kubenswrapper[4931]: E1129 00:00:45.029539 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs podName:5e38ad69-f5fb-415d-ad76-e899aed032a6 nodeName:}" failed. No retries permitted until 2025-11-29 00:01:01.029507175 +0000 UTC m=+74.191400447 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs") pod "network-metrics-daemon-7dwtb" (UID: "5e38ad69-f5fb-415d-ad76-e899aed032a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.081662 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.081714 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.081727 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.081747 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.081762 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:45Z","lastTransitionTime":"2025-11-29T00:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.184490 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.184916 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.184933 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.184962 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.184983 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:45Z","lastTransitionTime":"2025-11-29T00:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.211264 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.211331 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:45 crc kubenswrapper[4931]: E1129 00:00:45.211472 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.211536 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:45 crc kubenswrapper[4931]: E1129 00:00:45.211690 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:45 crc kubenswrapper[4931]: E1129 00:00:45.211799 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.288651 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.288725 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.288776 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.288839 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.288861 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:45Z","lastTransitionTime":"2025-11-29T00:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.392433 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.392507 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.392527 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.392555 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.392578 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:45Z","lastTransitionTime":"2025-11-29T00:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.496482 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.496567 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.496588 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.496622 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.496746 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:45Z","lastTransitionTime":"2025-11-29T00:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.600466 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.600554 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.600619 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.600654 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.600674 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:45Z","lastTransitionTime":"2025-11-29T00:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.703524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.703594 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.703616 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.703647 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.703668 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:45Z","lastTransitionTime":"2025-11-29T00:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.808018 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.808112 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.808136 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.808172 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.808195 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:45Z","lastTransitionTime":"2025-11-29T00:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.912146 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.912211 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.912234 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.912262 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:45 crc kubenswrapper[4931]: I1129 00:00:45.912283 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:45Z","lastTransitionTime":"2025-11-29T00:00:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.016077 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.016186 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.016205 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.016232 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.016252 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:46Z","lastTransitionTime":"2025-11-29T00:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.121139 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.122549 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.122733 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.123001 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.123200 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:46Z","lastTransitionTime":"2025-11-29T00:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.211862 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:46 crc kubenswrapper[4931]: E1129 00:00:46.212093 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.227208 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.227291 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.227319 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.227353 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.227379 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:46Z","lastTransitionTime":"2025-11-29T00:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.331096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.331169 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.331188 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.331218 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.331238 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:46Z","lastTransitionTime":"2025-11-29T00:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.434999 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.435078 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.435104 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.435145 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.435174 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:46Z","lastTransitionTime":"2025-11-29T00:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.538885 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.538959 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.538979 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.539013 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.539035 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:46Z","lastTransitionTime":"2025-11-29T00:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.642903 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.642971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.642991 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.643021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.643041 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:46Z","lastTransitionTime":"2025-11-29T00:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.746239 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.746305 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.746324 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.746354 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.746372 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:46Z","lastTransitionTime":"2025-11-29T00:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.849398 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.849475 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.849494 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.849523 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.849541 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:46Z","lastTransitionTime":"2025-11-29T00:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.953088 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.953159 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.953178 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.953206 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:46 crc kubenswrapper[4931]: I1129 00:00:46.953224 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:46Z","lastTransitionTime":"2025-11-29T00:00:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.056928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.056983 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.057001 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.057025 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.057042 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:47Z","lastTransitionTime":"2025-11-29T00:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.159413 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.159452 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.159463 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.159481 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.159493 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:47Z","lastTransitionTime":"2025-11-29T00:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.211276 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.211369 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.211387 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:47 crc kubenswrapper[4931]: E1129 00:00:47.211441 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:47 crc kubenswrapper[4931]: E1129 00:00:47.211620 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:47 crc kubenswrapper[4931]: E1129 00:00:47.211772 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.230395 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.250841 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.261748 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.261861 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.261892 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.261926 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.261951 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:47Z","lastTransitionTime":"2025-11-29T00:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.274210 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.293933 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.308655 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.327000 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00e8b190-9f98-45be-9254-2961ad8064a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c38778155a771c99ed4c71c12be76871dc3299129f77a1551fc19f8482e4065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9dc7cc6f5295935a3444f30c164cc63f6b2639a70fd4f45098f28ae9d0f98d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b89687792ce16c3406e8500eb2bfa867fcdc1db444a0623caca0b5033229f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.347964 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.365109 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.365453 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.365555 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.365578 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.365641 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.365663 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:47Z","lastTransitionTime":"2025-11-29T00:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.389436 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.407832 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.432444 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.455236 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.471611 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.471693 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.471751 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.472135 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.472168 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:47Z","lastTransitionTime":"2025-11-29T00:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.489520 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:40Z\\\",\\\"message\\\":\\\" 6599 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:40.145441 6599 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:00:40.145452 6599 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:40.145480 6599 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 00:00:40.145520 6599 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:40.145531 6599 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 00:00:40.145539 6599 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 00:00:40.145571 6599 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:40.145630 6599 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:00:40.145668 6599 factory.go:656] Stopping watch factory\\\\nI1129 00:00:40.145692 6599 ovnkube.go:599] Stopped ovnkube\\\\nI1129 00:00:40.145728 6599 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:00:40.145738 6599 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 00:00:40.145747 6599 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 00:00:40.145792 6599 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF1129 00:00:40.145877 6599 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.525360 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.541879 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.561621 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.574537 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.574626 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.574647 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.574711 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.574733 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:47Z","lastTransitionTime":"2025-11-29T00:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.582459 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.606158 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:47Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.677885 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.678258 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.678405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.678554 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.678719 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:47Z","lastTransitionTime":"2025-11-29T00:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.782490 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.782539 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.782556 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.782580 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.782598 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:47Z","lastTransitionTime":"2025-11-29T00:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.886281 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.886336 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.886352 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.886373 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.886386 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:47Z","lastTransitionTime":"2025-11-29T00:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.990099 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.990177 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.990197 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.990225 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:47 crc kubenswrapper[4931]: I1129 00:00:47.990249 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:47Z","lastTransitionTime":"2025-11-29T00:00:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.093776 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.093855 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.093870 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.093889 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.093903 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:48Z","lastTransitionTime":"2025-11-29T00:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.196870 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.196936 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.196992 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.197028 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.197047 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:48Z","lastTransitionTime":"2025-11-29T00:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.211654 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:48 crc kubenswrapper[4931]: E1129 00:00:48.211877 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.300015 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.300052 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.300061 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.300076 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.300087 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:48Z","lastTransitionTime":"2025-11-29T00:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.402840 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.402882 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.402891 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.402910 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.402923 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:48Z","lastTransitionTime":"2025-11-29T00:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.505352 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.505413 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.505431 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.505462 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.505481 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:48Z","lastTransitionTime":"2025-11-29T00:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.608860 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.608918 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.608933 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.608954 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.608971 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:48Z","lastTransitionTime":"2025-11-29T00:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.711869 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.711912 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.711923 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.711939 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.711949 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:48Z","lastTransitionTime":"2025-11-29T00:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.814613 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.814675 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.814693 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.814717 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.814736 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:48Z","lastTransitionTime":"2025-11-29T00:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.918021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.918107 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.918125 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.918634 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:48 crc kubenswrapper[4931]: I1129 00:00:48.918696 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:48Z","lastTransitionTime":"2025-11-29T00:00:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.021965 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.022496 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.022925 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.023165 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.023344 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:49Z","lastTransitionTime":"2025-11-29T00:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.126641 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.127106 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.127263 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.127449 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.127586 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:49Z","lastTransitionTime":"2025-11-29T00:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.211887 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.211982 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:49 crc kubenswrapper[4931]: E1129 00:00:49.212090 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:49 crc kubenswrapper[4931]: E1129 00:00:49.212188 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.211889 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:49 crc kubenswrapper[4931]: E1129 00:00:49.212340 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.230945 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.230992 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.231008 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.231029 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.231047 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:49Z","lastTransitionTime":"2025-11-29T00:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.334167 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.334218 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.334235 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.334260 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.334276 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:49Z","lastTransitionTime":"2025-11-29T00:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.437681 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.437729 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.437746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.437770 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.437787 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:49Z","lastTransitionTime":"2025-11-29T00:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.541423 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.541487 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.541505 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.541530 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.541549 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:49Z","lastTransitionTime":"2025-11-29T00:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.643944 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.644301 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.644385 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.644484 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.644581 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:49Z","lastTransitionTime":"2025-11-29T00:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.747173 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.747234 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.747250 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.747275 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.747295 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:49Z","lastTransitionTime":"2025-11-29T00:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.850533 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.851019 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.851210 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.851389 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.851611 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:49Z","lastTransitionTime":"2025-11-29T00:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.954599 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.954652 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.954662 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.954680 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:49 crc kubenswrapper[4931]: I1129 00:00:49.954694 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:49Z","lastTransitionTime":"2025-11-29T00:00:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.057983 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.058299 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.058555 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.058714 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.058887 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:50Z","lastTransitionTime":"2025-11-29T00:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.161831 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.161879 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.161889 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.161906 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.161918 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:50Z","lastTransitionTime":"2025-11-29T00:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.212024 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:50 crc kubenswrapper[4931]: E1129 00:00:50.212216 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.265298 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.265359 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.265412 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.265440 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.265460 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:50Z","lastTransitionTime":"2025-11-29T00:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.367615 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.367704 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.367728 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.367760 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.367782 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:50Z","lastTransitionTime":"2025-11-29T00:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.470602 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.470650 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.470663 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.470680 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.470694 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:50Z","lastTransitionTime":"2025-11-29T00:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.573214 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.573287 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.573306 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.573336 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.573356 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:50Z","lastTransitionTime":"2025-11-29T00:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.675605 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.675676 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.675695 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.675722 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.675756 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:50Z","lastTransitionTime":"2025-11-29T00:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.778452 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.778531 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.778563 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.778593 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.778613 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:50Z","lastTransitionTime":"2025-11-29T00:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.881695 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.881760 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.881778 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.881835 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.881863 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:50Z","lastTransitionTime":"2025-11-29T00:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.984602 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.984682 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.984701 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.984729 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:50 crc kubenswrapper[4931]: I1129 00:00:50.984758 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:50Z","lastTransitionTime":"2025-11-29T00:00:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.087892 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.087944 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.087957 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.087976 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.087988 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.191491 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.191565 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.191583 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.191611 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.191630 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.211670 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.211729 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.211734 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:51 crc kubenswrapper[4931]: E1129 00:00:51.211905 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:51 crc kubenswrapper[4931]: E1129 00:00:51.212029 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:51 crc kubenswrapper[4931]: E1129 00:00:51.212160 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.294508 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.294581 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.294591 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.294608 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.294619 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.397970 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.398042 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.398063 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.398097 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.398128 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.502527 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.502617 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.502638 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.502671 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.502693 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.605735 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.605837 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.605855 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.605877 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.605892 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.708141 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.708192 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.708202 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.708222 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.708234 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.810669 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.810749 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.810768 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.810801 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.810846 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.824497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.824543 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.824560 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.824583 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.824601 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: E1129 00:00:51.840546 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:51Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.845176 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.845234 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.845251 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.845275 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.845293 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: E1129 00:00:51.861868 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:51Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.865801 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.865891 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.865913 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.865945 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.865966 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: E1129 00:00:51.878963 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:51Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.884826 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.884874 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.884884 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.885242 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.885266 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: E1129 00:00:51.898979 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:51Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.903108 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.903170 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.903182 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.903204 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.903219 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:51 crc kubenswrapper[4931]: E1129 00:00:51.916043 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:51Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:51 crc kubenswrapper[4931]: E1129 00:00:51.916207 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.918082 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.918171 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.918187 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.918212 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:51 crc kubenswrapper[4931]: I1129 00:00:51.918224 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:51Z","lastTransitionTime":"2025-11-29T00:00:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.022269 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.022310 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.022346 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.022367 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.022379 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:52Z","lastTransitionTime":"2025-11-29T00:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.125052 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.125086 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.125094 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.125108 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.125119 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:52Z","lastTransitionTime":"2025-11-29T00:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.211509 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:52 crc kubenswrapper[4931]: E1129 00:00:52.211754 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.228599 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.228653 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.228672 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.228706 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.228727 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:52Z","lastTransitionTime":"2025-11-29T00:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.331659 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.331710 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.331727 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.331753 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.331771 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:52Z","lastTransitionTime":"2025-11-29T00:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.435366 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.435441 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.435459 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.435490 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.435511 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:52Z","lastTransitionTime":"2025-11-29T00:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.538799 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.539072 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.539096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.539127 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.539147 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:52Z","lastTransitionTime":"2025-11-29T00:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.641753 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.641877 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.641910 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.641950 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.641978 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:52Z","lastTransitionTime":"2025-11-29T00:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.745042 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.745115 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.745133 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.745164 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.745183 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:52Z","lastTransitionTime":"2025-11-29T00:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.849269 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.849328 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.849347 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.849376 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.849396 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:52Z","lastTransitionTime":"2025-11-29T00:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.952191 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.952258 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.952269 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.952291 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:52 crc kubenswrapper[4931]: I1129 00:00:52.952304 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:52Z","lastTransitionTime":"2025-11-29T00:00:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.055258 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.055598 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.055709 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.055821 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.055920 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:53Z","lastTransitionTime":"2025-11-29T00:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.159027 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.159092 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.159107 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.159128 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.159141 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:53Z","lastTransitionTime":"2025-11-29T00:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.212126 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.212260 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:53 crc kubenswrapper[4931]: E1129 00:00:53.212304 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.212132 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:53 crc kubenswrapper[4931]: E1129 00:00:53.212478 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:53 crc kubenswrapper[4931]: E1129 00:00:53.212660 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.262133 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.262181 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.262198 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.262222 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.262236 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:53Z","lastTransitionTime":"2025-11-29T00:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.366100 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.366158 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.366171 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.366194 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.366210 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:53Z","lastTransitionTime":"2025-11-29T00:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.469039 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.469117 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.469136 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.469191 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.469214 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:53Z","lastTransitionTime":"2025-11-29T00:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.572490 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.572538 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.572547 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.572561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.572573 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:53Z","lastTransitionTime":"2025-11-29T00:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.674919 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.674993 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.675017 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.675052 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.675074 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:53Z","lastTransitionTime":"2025-11-29T00:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.778436 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.778492 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.778512 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.778537 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.778555 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:53Z","lastTransitionTime":"2025-11-29T00:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.881292 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.881354 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.881382 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.881411 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.881435 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:53Z","lastTransitionTime":"2025-11-29T00:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.985791 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.985844 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.985854 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.985871 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:53 crc kubenswrapper[4931]: I1129 00:00:53.985884 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:53Z","lastTransitionTime":"2025-11-29T00:00:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.088705 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.089159 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.089258 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.089329 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.089394 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:54Z","lastTransitionTime":"2025-11-29T00:00:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.192628 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.192716 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.192735 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.192767 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.192786 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:54Z","lastTransitionTime":"2025-11-29T00:00:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.211432 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:54 crc kubenswrapper[4931]: E1129 00:00:54.211969 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.212184 4931 scope.go:117] "RemoveContainer" containerID="32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232" Nov 29 00:00:54 crc kubenswrapper[4931]: E1129 00:00:54.212449 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.296026 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.296101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.296118 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.296146 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.296167 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:54Z","lastTransitionTime":"2025-11-29T00:00:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.399186 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.399242 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.399255 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.399276 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.399290 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:54Z","lastTransitionTime":"2025-11-29T00:00:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.501558 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.501595 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.501606 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.501621 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.501633 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:54Z","lastTransitionTime":"2025-11-29T00:00:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.604539 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.604568 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.604577 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.604590 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.604599 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:54Z","lastTransitionTime":"2025-11-29T00:00:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.706787 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.706860 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.706879 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.706900 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.706915 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:54Z","lastTransitionTime":"2025-11-29T00:00:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.810302 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.810353 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.810366 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.810385 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.810400 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:54Z","lastTransitionTime":"2025-11-29T00:00:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.912745 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.912789 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.912800 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.912844 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:54 crc kubenswrapper[4931]: I1129 00:00:54.912859 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:54Z","lastTransitionTime":"2025-11-29T00:00:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.015388 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.015491 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.015505 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.015531 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.015546 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:55Z","lastTransitionTime":"2025-11-29T00:00:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.118643 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.118707 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.118724 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.118751 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.118768 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:55Z","lastTransitionTime":"2025-11-29T00:00:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.211496 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.211519 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.211708 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:55 crc kubenswrapper[4931]: E1129 00:00:55.211917 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:55 crc kubenswrapper[4931]: E1129 00:00:55.212638 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:55 crc kubenswrapper[4931]: E1129 00:00:55.212751 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.221481 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.221539 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.221563 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.221595 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.221615 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:55Z","lastTransitionTime":"2025-11-29T00:00:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.227206 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.323903 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.323939 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.323951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.323968 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.324000 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:55Z","lastTransitionTime":"2025-11-29T00:00:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.427346 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.427755 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.427770 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.427790 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.427825 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:55Z","lastTransitionTime":"2025-11-29T00:00:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.531423 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.531496 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.531515 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.531550 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.531569 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:55Z","lastTransitionTime":"2025-11-29T00:00:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.635094 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.635139 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.635151 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.635170 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.635183 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:55Z","lastTransitionTime":"2025-11-29T00:00:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.739495 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.739544 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.739555 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.739574 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.739586 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:55Z","lastTransitionTime":"2025-11-29T00:00:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.842303 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.842367 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.842387 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.842413 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.842429 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:55Z","lastTransitionTime":"2025-11-29T00:00:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.945957 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.946036 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.946059 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.946090 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:55 crc kubenswrapper[4931]: I1129 00:00:55.946113 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:55Z","lastTransitionTime":"2025-11-29T00:00:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.049871 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.049947 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.049971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.050009 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.050039 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:56Z","lastTransitionTime":"2025-11-29T00:00:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.153245 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.153334 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.153357 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.153389 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.153414 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:56Z","lastTransitionTime":"2025-11-29T00:00:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.211269 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:56 crc kubenswrapper[4931]: E1129 00:00:56.211480 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.256402 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.256495 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.256519 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.256554 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.256575 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:56Z","lastTransitionTime":"2025-11-29T00:00:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.358870 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.358929 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.358947 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.358972 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.358993 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:56Z","lastTransitionTime":"2025-11-29T00:00:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.461439 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.461501 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.461521 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.461548 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.461567 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:56Z","lastTransitionTime":"2025-11-29T00:00:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.564278 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.564368 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.564392 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.564430 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.564453 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:56Z","lastTransitionTime":"2025-11-29T00:00:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.667586 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.667648 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.667661 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.667683 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.667698 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:56Z","lastTransitionTime":"2025-11-29T00:00:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.770788 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.770877 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.770892 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.770919 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.770938 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:56Z","lastTransitionTime":"2025-11-29T00:00:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.873029 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.873088 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.873107 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.873134 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.873155 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:56Z","lastTransitionTime":"2025-11-29T00:00:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.975742 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.975781 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.975792 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.975828 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:56 crc kubenswrapper[4931]: I1129 00:00:56.975840 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:56Z","lastTransitionTime":"2025-11-29T00:00:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.078261 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.078312 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.078326 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.078351 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.078364 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:57Z","lastTransitionTime":"2025-11-29T00:00:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.181314 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.181370 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.181383 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.181406 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.181423 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:57Z","lastTransitionTime":"2025-11-29T00:00:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.211362 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.211417 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:57 crc kubenswrapper[4931]: E1129 00:00:57.211669 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.211701 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:57 crc kubenswrapper[4931]: E1129 00:00:57.211906 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:57 crc kubenswrapper[4931]: E1129 00:00:57.212072 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.227720 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.241897 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00e8b190-9f98-45be-9254-2961ad8064a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c38778155a771c99ed4c71c12be76871dc3299129f77a1551fc19f8482e4065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9dc7cc6f5295935a3444f30c164cc63f6b2639a70fd4f45098f28ae9d0f98d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b89687792ce16c3406e8500eb2bfa867fcdc1db444a0623caca0b5033229f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.254527 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7111f8fe-6cd1-4dc5-99fb-2ce20e485267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53bafde2aa1cbcfcdb36a53eabbf234b09505c13a764ae7ef41659856767dfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.268607 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.280902 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.284921 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.284974 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.284990 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.285017 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.285034 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:57Z","lastTransitionTime":"2025-11-29T00:00:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.294685 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.311168 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.326086 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.368214 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:40Z\\\",\\\"message\\\":\\\" 6599 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:40.145441 6599 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:00:40.145452 6599 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:40.145480 6599 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 00:00:40.145520 6599 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:40.145531 6599 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 00:00:40.145539 6599 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 00:00:40.145571 6599 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:40.145630 6599 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:00:40.145668 6599 factory.go:656] Stopping watch factory\\\\nI1129 00:00:40.145692 6599 ovnkube.go:599] Stopped ovnkube\\\\nI1129 00:00:40.145728 6599 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:00:40.145738 6599 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 00:00:40.145747 6599 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 00:00:40.145792 6599 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF1129 00:00:40.145877 6599 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.388381 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.388413 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.388422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.388438 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.388448 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:57Z","lastTransitionTime":"2025-11-29T00:00:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.402111 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.422575 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.441377 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.461139 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.479927 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.493984 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.494034 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.494046 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.494071 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.494087 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:57Z","lastTransitionTime":"2025-11-29T00:00:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.502096 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.519028 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.534470 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.552052 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.567321 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:00:57Z is after 2025-08-24T17:21:41Z" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.595902 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.595971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.595990 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.596016 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.596035 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:57Z","lastTransitionTime":"2025-11-29T00:00:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.698588 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.698650 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.698668 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.698695 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.698718 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:57Z","lastTransitionTime":"2025-11-29T00:00:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.801209 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.801273 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.801292 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.801322 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.801338 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:57Z","lastTransitionTime":"2025-11-29T00:00:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.904367 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.904432 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.904457 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.904494 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:57 crc kubenswrapper[4931]: I1129 00:00:57.904519 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:57Z","lastTransitionTime":"2025-11-29T00:00:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.006847 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.006893 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.006902 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.006922 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.006933 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:58Z","lastTransitionTime":"2025-11-29T00:00:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.109583 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.109638 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.109647 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.109664 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.109675 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:58Z","lastTransitionTime":"2025-11-29T00:00:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.211459 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:00:58 crc kubenswrapper[4931]: E1129 00:00:58.211619 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.212952 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.212993 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.213006 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.213024 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.213038 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:58Z","lastTransitionTime":"2025-11-29T00:00:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.315851 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.315891 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.315904 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.315927 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.315941 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:58Z","lastTransitionTime":"2025-11-29T00:00:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.418756 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.418833 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.418850 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.418871 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.418886 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:58Z","lastTransitionTime":"2025-11-29T00:00:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.521951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.522026 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.522045 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.522076 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.522096 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:58Z","lastTransitionTime":"2025-11-29T00:00:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.625363 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.625419 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.625437 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.625464 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.625484 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:58Z","lastTransitionTime":"2025-11-29T00:00:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.728556 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.728630 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.728647 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.728674 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.728691 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:58Z","lastTransitionTime":"2025-11-29T00:00:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.831454 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.831488 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.831495 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.831509 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.831519 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:58Z","lastTransitionTime":"2025-11-29T00:00:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.934461 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.934531 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.934552 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.934582 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:58 crc kubenswrapper[4931]: I1129 00:00:58.934604 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:58Z","lastTransitionTime":"2025-11-29T00:00:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.037229 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.037275 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.037286 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.037305 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.037318 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:59Z","lastTransitionTime":"2025-11-29T00:00:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.140666 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.140735 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.140761 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.140795 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.140869 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:59Z","lastTransitionTime":"2025-11-29T00:00:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.211454 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.211500 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:00:59 crc kubenswrapper[4931]: E1129 00:00:59.211654 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.211736 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:00:59 crc kubenswrapper[4931]: E1129 00:00:59.211948 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:00:59 crc kubenswrapper[4931]: E1129 00:00:59.212081 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.243623 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.243710 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.243732 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.243765 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.243791 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:59Z","lastTransitionTime":"2025-11-29T00:00:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.345729 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.345786 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.345799 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.345834 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.345849 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:59Z","lastTransitionTime":"2025-11-29T00:00:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.447518 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.447557 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.447566 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.447584 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.447595 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:59Z","lastTransitionTime":"2025-11-29T00:00:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.551322 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.551369 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.551379 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.551397 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.551406 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:59Z","lastTransitionTime":"2025-11-29T00:00:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.653938 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.653986 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.653998 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.654019 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.654031 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:59Z","lastTransitionTime":"2025-11-29T00:00:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.756997 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.757059 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.757077 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.757105 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.757123 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:59Z","lastTransitionTime":"2025-11-29T00:00:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.860197 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.860272 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.860288 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.860314 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.860332 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:59Z","lastTransitionTime":"2025-11-29T00:00:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.963576 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.963656 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.963677 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.963709 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:00:59 crc kubenswrapper[4931]: I1129 00:00:59.963732 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:00:59Z","lastTransitionTime":"2025-11-29T00:00:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.072578 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.072650 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.072674 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.072706 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.072729 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:00Z","lastTransitionTime":"2025-11-29T00:01:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.175624 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.175715 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.175746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.175785 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.175847 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:00Z","lastTransitionTime":"2025-11-29T00:01:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.212039 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:00 crc kubenswrapper[4931]: E1129 00:01:00.212169 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.279902 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.279967 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.279983 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.280011 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.280029 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:00Z","lastTransitionTime":"2025-11-29T00:01:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.382943 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.383008 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.383027 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.383055 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.383077 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:00Z","lastTransitionTime":"2025-11-29T00:01:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.486137 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.486200 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.486219 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.486246 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.486267 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:00Z","lastTransitionTime":"2025-11-29T00:01:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.589382 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.589483 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.589502 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.589530 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.589549 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:00Z","lastTransitionTime":"2025-11-29T00:01:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.692657 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.692726 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.692746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.692773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.692795 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:00Z","lastTransitionTime":"2025-11-29T00:01:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.795622 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.795688 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.795705 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.795736 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.795755 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:00Z","lastTransitionTime":"2025-11-29T00:01:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.898197 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.898241 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.898252 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.898273 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:00 crc kubenswrapper[4931]: I1129 00:01:00.898288 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:00Z","lastTransitionTime":"2025-11-29T00:01:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.001438 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.001503 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.001521 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.001591 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.001611 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:01Z","lastTransitionTime":"2025-11-29T00:01:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.105301 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.105347 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.105359 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.105380 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.105394 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:01Z","lastTransitionTime":"2025-11-29T00:01:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.109087 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:01 crc kubenswrapper[4931]: E1129 00:01:01.109266 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:01:01 crc kubenswrapper[4931]: E1129 00:01:01.109378 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs podName:5e38ad69-f5fb-415d-ad76-e899aed032a6 nodeName:}" failed. No retries permitted until 2025-11-29 00:01:33.109309369 +0000 UTC m=+106.271202611 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs") pod "network-metrics-daemon-7dwtb" (UID: "5e38ad69-f5fb-415d-ad76-e899aed032a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.209328 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.209497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.209534 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.209607 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.209630 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:01Z","lastTransitionTime":"2025-11-29T00:01:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.211596 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.211782 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:01 crc kubenswrapper[4931]: E1129 00:01:01.211959 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.212012 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:01 crc kubenswrapper[4931]: E1129 00:01:01.212313 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:01 crc kubenswrapper[4931]: E1129 00:01:01.212415 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.313283 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.313553 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.313709 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.313922 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.314078 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:01Z","lastTransitionTime":"2025-11-29T00:01:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.416770 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.416903 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.416926 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.416956 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.416974 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:01Z","lastTransitionTime":"2025-11-29T00:01:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.520723 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.521131 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.521328 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.521489 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.521709 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:01Z","lastTransitionTime":"2025-11-29T00:01:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.624640 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.624685 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.624696 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.624717 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.624728 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:01Z","lastTransitionTime":"2025-11-29T00:01:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.728512 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.729081 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.729239 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.729449 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.729596 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:01Z","lastTransitionTime":"2025-11-29T00:01:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.738363 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22vkg_31e6835b-462a-4a4f-9ba9-74531c14d41a/kube-multus/0.log" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.738444 4931 generic.go:334] "Generic (PLEG): container finished" podID="31e6835b-462a-4a4f-9ba9-74531c14d41a" containerID="eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857" exitCode=1 Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.738489 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22vkg" event={"ID":"31e6835b-462a-4a4f-9ba9-74531c14d41a","Type":"ContainerDied","Data":"eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857"} Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.739036 4931 scope.go:117] "RemoveContainer" containerID="eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.757854 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7111f8fe-6cd1-4dc5-99fb-2ce20e485267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53bafde2aa1cbcfcdb36a53eabbf234b09505c13a764ae7ef41659856767dfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.779282 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.801217 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.818162 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.833488 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.833886 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.834048 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.834201 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.834363 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:01Z","lastTransitionTime":"2025-11-29T00:01:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.836662 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00e8b190-9f98-45be-9254-2961ad8064a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c38778155a771c99ed4c71c12be76871dc3299129f77a1551fc19f8482e4065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9dc7cc6f5295935a3444f30c164cc63f6b2639a70fd4f45098f28ae9d0f98d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b89687792ce16c3406e8500eb2bfa867fcdc1db444a0623caca0b5033229f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.852560 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.870081 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.885950 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.903224 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.922121 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.937319 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.937361 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.937372 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.937391 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.937403 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:01Z","lastTransitionTime":"2025-11-29T00:01:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.944379 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:01:01Z\\\",\\\"message\\\":\\\"2025-11-29T00:00:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d\\\\n2025-11-29T00:00:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d to /host/opt/cni/bin/\\\\n2025-11-29T00:00:16Z [verbose] multus-daemon started\\\\n2025-11-29T00:00:16Z [verbose] Readiness Indicator file check\\\\n2025-11-29T00:01:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:01 crc kubenswrapper[4931]: I1129 00:01:01.975367 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:40Z\\\",\\\"message\\\":\\\" 6599 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:40.145441 6599 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:00:40.145452 6599 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:40.145480 6599 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 00:00:40.145520 6599 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:40.145531 6599 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 00:00:40.145539 6599 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 00:00:40.145571 6599 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:40.145630 6599 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:00:40.145668 6599 factory.go:656] Stopping watch factory\\\\nI1129 00:00:40.145692 6599 ovnkube.go:599] Stopped ovnkube\\\\nI1129 00:00:40.145728 6599 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:00:40.145738 6599 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 00:00:40.145747 6599 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 00:00:40.145792 6599 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF1129 00:00:40.145877 6599 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.000556 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:01Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.016586 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.033619 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.040682 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.040715 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.040756 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.040779 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.040793 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.049185 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.070478 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.086219 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.098355 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.098405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.098447 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.098466 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.098478 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.103308 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: E1129 00:01:02.110824 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.114149 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.114211 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.114224 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.114244 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.114261 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: E1129 00:01:02.126778 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.131178 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.131266 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.131294 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.131334 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.131359 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: E1129 00:01:02.146194 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.151029 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.151086 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.151099 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.151118 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.151148 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: E1129 00:01:02.164201 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.167989 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.168022 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.168033 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.168053 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.168068 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: E1129 00:01:02.183057 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: E1129 00:01:02.183215 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.185070 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.185101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.185113 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.185133 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.185149 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.212038 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:02 crc kubenswrapper[4931]: E1129 00:01:02.212198 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.287754 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.287788 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.287799 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.287844 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.287854 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.390005 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.390059 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.390082 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.390107 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.390123 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.493038 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.493087 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.493104 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.493129 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.493146 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.595837 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.595910 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.595929 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.595959 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.595981 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.699172 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.699246 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.699264 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.699293 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.699314 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.745944 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22vkg_31e6835b-462a-4a4f-9ba9-74531c14d41a/kube-multus/0.log" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.746038 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22vkg" event={"ID":"31e6835b-462a-4a4f-9ba9-74531c14d41a","Type":"ContainerStarted","Data":"7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652"} Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.771333 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.796278 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.802271 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.802360 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.802398 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.802422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.802436 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.817674 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.839851 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.855346 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.875620 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00e8b190-9f98-45be-9254-2961ad8064a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c38778155a771c99ed4c71c12be76871dc3299129f77a1551fc19f8482e4065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9dc7cc6f5295935a3444f30c164cc63f6b2639a70fd4f45098f28ae9d0f98d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b89687792ce16c3406e8500eb2bfa867fcdc1db444a0623caca0b5033229f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.893630 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7111f8fe-6cd1-4dc5-99fb-2ce20e485267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53bafde2aa1cbcfcdb36a53eabbf234b09505c13a764ae7ef41659856767dfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.905405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.905476 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.905496 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.905538 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.905565 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:02Z","lastTransitionTime":"2025-11-29T00:01:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.920406 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.945723 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.960644 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:02 crc kubenswrapper[4931]: I1129 00:01:02.991196 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:02Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.008417 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:03Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.008946 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.009031 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.009052 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.009080 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.009100 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:03Z","lastTransitionTime":"2025-11-29T00:01:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.030605 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:03Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.047560 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:03Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.062504 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:03Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.089597 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:03Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.107349 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:01:01Z\\\",\\\"message\\\":\\\"2025-11-29T00:00:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d\\\\n2025-11-29T00:00:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d to /host/opt/cni/bin/\\\\n2025-11-29T00:00:16Z [verbose] multus-daemon started\\\\n2025-11-29T00:00:16Z [verbose] Readiness Indicator file check\\\\n2025-11-29T00:01:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:03Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.114018 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.114074 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.114088 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.114108 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.114122 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:03Z","lastTransitionTime":"2025-11-29T00:01:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.141063 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:40Z\\\",\\\"message\\\":\\\" 6599 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:40.145441 6599 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:00:40.145452 6599 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:40.145480 6599 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 00:00:40.145520 6599 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:40.145531 6599 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 00:00:40.145539 6599 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 00:00:40.145571 6599 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:40.145630 6599 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:00:40.145668 6599 factory.go:656] Stopping watch factory\\\\nI1129 00:00:40.145692 6599 ovnkube.go:599] Stopped ovnkube\\\\nI1129 00:00:40.145728 6599 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:00:40.145738 6599 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 00:00:40.145747 6599 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 00:00:40.145792 6599 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF1129 00:00:40.145877 6599 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:03Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.157381 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:03Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.212284 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.212378 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.212476 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:03 crc kubenswrapper[4931]: E1129 00:01:03.212582 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:03 crc kubenswrapper[4931]: E1129 00:01:03.212670 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:03 crc kubenswrapper[4931]: E1129 00:01:03.212779 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.217626 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.217665 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.217676 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.217696 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.217711 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:03Z","lastTransitionTime":"2025-11-29T00:01:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.320376 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.320428 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.320439 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.320459 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.320472 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:03Z","lastTransitionTime":"2025-11-29T00:01:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.424130 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.424196 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.424215 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.424245 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.424266 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:03Z","lastTransitionTime":"2025-11-29T00:01:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.526683 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.526761 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.526777 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.526840 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.526856 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:03Z","lastTransitionTime":"2025-11-29T00:01:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.630509 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.630568 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.630580 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.630605 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.630618 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:03Z","lastTransitionTime":"2025-11-29T00:01:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.733531 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.733578 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.733589 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.733607 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.733620 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:03Z","lastTransitionTime":"2025-11-29T00:01:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.836692 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.836751 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.836767 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.836790 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.836805 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:03Z","lastTransitionTime":"2025-11-29T00:01:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.939952 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.939999 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.940009 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.940029 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:03 crc kubenswrapper[4931]: I1129 00:01:03.940043 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:03Z","lastTransitionTime":"2025-11-29T00:01:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.043459 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.043524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.043536 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.043562 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.043582 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:04Z","lastTransitionTime":"2025-11-29T00:01:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.146967 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.147274 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.147286 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.147307 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.147319 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:04Z","lastTransitionTime":"2025-11-29T00:01:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.211978 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:04 crc kubenswrapper[4931]: E1129 00:01:04.212169 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.250768 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.250880 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.250920 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.250957 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.250983 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:04Z","lastTransitionTime":"2025-11-29T00:01:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.354303 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.354376 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.354397 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.354429 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.354454 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:04Z","lastTransitionTime":"2025-11-29T00:01:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.457918 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.458049 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.458073 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.458097 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.458115 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:04Z","lastTransitionTime":"2025-11-29T00:01:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.561364 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.561445 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.561466 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.561496 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.561518 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:04Z","lastTransitionTime":"2025-11-29T00:01:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.664632 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.664711 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.664737 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.664774 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.664799 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:04Z","lastTransitionTime":"2025-11-29T00:01:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.768295 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.768351 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.768372 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.768399 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.768419 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:04Z","lastTransitionTime":"2025-11-29T00:01:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.875179 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.875252 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.875271 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.875298 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.875319 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:04Z","lastTransitionTime":"2025-11-29T00:01:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.979622 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.979691 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.979709 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.979737 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:04 crc kubenswrapper[4931]: I1129 00:01:04.979762 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:04Z","lastTransitionTime":"2025-11-29T00:01:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.082131 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.082173 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.082186 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.082231 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.082248 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:05Z","lastTransitionTime":"2025-11-29T00:01:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.185245 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.185317 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.185342 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.185375 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.185401 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:05Z","lastTransitionTime":"2025-11-29T00:01:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.213062 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.213308 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:05 crc kubenswrapper[4931]: E1129 00:01:05.213433 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.213487 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:05 crc kubenswrapper[4931]: E1129 00:01:05.213694 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:05 crc kubenswrapper[4931]: E1129 00:01:05.213782 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.287843 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.287910 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.287929 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.287953 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.287972 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:05Z","lastTransitionTime":"2025-11-29T00:01:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.390875 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.390917 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.390929 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.390949 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.390963 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:05Z","lastTransitionTime":"2025-11-29T00:01:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.493661 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.493703 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.493715 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.493736 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.493749 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:05Z","lastTransitionTime":"2025-11-29T00:01:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.596788 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.596918 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.596938 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.596965 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.596983 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:05Z","lastTransitionTime":"2025-11-29T00:01:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.699922 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.699995 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.700021 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.700055 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.700081 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:05Z","lastTransitionTime":"2025-11-29T00:01:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.802842 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.802960 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.802978 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.802999 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.803013 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:05Z","lastTransitionTime":"2025-11-29T00:01:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.905713 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.905794 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.905875 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.905912 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:05 crc kubenswrapper[4931]: I1129 00:01:05.905941 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:05Z","lastTransitionTime":"2025-11-29T00:01:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.009285 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.009337 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.009349 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.009370 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.009385 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:06Z","lastTransitionTime":"2025-11-29T00:01:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.111887 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.111958 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.111984 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.112020 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.112039 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:06Z","lastTransitionTime":"2025-11-29T00:01:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.211962 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:06 crc kubenswrapper[4931]: E1129 00:01:06.212117 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.215026 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.215087 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.215142 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.215171 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.215198 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:06Z","lastTransitionTime":"2025-11-29T00:01:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.318392 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.318454 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.318470 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.318502 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.318521 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:06Z","lastTransitionTime":"2025-11-29T00:01:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.421468 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.421553 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.421578 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.421615 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.421640 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:06Z","lastTransitionTime":"2025-11-29T00:01:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.524145 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.524209 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.524226 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.524252 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.524274 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:06Z","lastTransitionTime":"2025-11-29T00:01:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.627665 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.627772 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.627854 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.627896 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.627924 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:06Z","lastTransitionTime":"2025-11-29T00:01:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.731219 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.731278 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.731296 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.731325 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.731345 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:06Z","lastTransitionTime":"2025-11-29T00:01:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.834916 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.834987 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.835006 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.835037 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.835059 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:06Z","lastTransitionTime":"2025-11-29T00:01:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.938070 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.938137 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.938161 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.938192 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:06 crc kubenswrapper[4931]: I1129 00:01:06.938214 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:06Z","lastTransitionTime":"2025-11-29T00:01:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.041124 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.041186 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.041197 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.041218 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.041229 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:07Z","lastTransitionTime":"2025-11-29T00:01:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.145045 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.145095 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.145104 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.145122 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.145136 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:07Z","lastTransitionTime":"2025-11-29T00:01:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.211416 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.211460 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:07 crc kubenswrapper[4931]: E1129 00:01:07.211628 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.212014 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:07 crc kubenswrapper[4931]: E1129 00:01:07.212122 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:07 crc kubenswrapper[4931]: E1129 00:01:07.212525 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.233565 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.248305 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.248353 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.248371 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.248396 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.248413 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:07Z","lastTransitionTime":"2025-11-29T00:01:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.249452 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.265610 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.287587 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.298988 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.316241 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00e8b190-9f98-45be-9254-2961ad8064a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c38778155a771c99ed4c71c12be76871dc3299129f77a1551fc19f8482e4065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9dc7cc6f5295935a3444f30c164cc63f6b2639a70fd4f45098f28ae9d0f98d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b89687792ce16c3406e8500eb2bfa867fcdc1db444a0623caca0b5033229f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.328731 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7111f8fe-6cd1-4dc5-99fb-2ce20e485267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53bafde2aa1cbcfcdb36a53eabbf234b09505c13a764ae7ef41659856767dfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.348568 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.352318 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.352626 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.352832 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.353024 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.353198 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:07Z","lastTransitionTime":"2025-11-29T00:01:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.366469 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.386992 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.408697 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:40Z\\\",\\\"message\\\":\\\" 6599 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:40.145441 6599 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:00:40.145452 6599 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:40.145480 6599 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 00:00:40.145520 6599 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:40.145531 6599 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 00:00:40.145539 6599 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 00:00:40.145571 6599 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:40.145630 6599 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:00:40.145668 6599 factory.go:656] Stopping watch factory\\\\nI1129 00:00:40.145692 6599 ovnkube.go:599] Stopped ovnkube\\\\nI1129 00:00:40.145728 6599 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:00:40.145738 6599 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 00:00:40.145747 6599 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 00:00:40.145792 6599 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF1129 00:00:40.145877 6599 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.442248 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.455715 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.455752 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.455760 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.455776 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.455787 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:07Z","lastTransitionTime":"2025-11-29T00:01:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.458155 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.473989 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.495284 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.509522 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.525179 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.539820 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:01:01Z\\\",\\\"message\\\":\\\"2025-11-29T00:00:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d\\\\n2025-11-29T00:00:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d to /host/opt/cni/bin/\\\\n2025-11-29T00:00:16Z [verbose] multus-daemon started\\\\n2025-11-29T00:00:16Z [verbose] Readiness Indicator file check\\\\n2025-11-29T00:01:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.550930 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:07Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.558700 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.558735 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.558749 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.558766 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.558780 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:07Z","lastTransitionTime":"2025-11-29T00:01:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.661925 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.662003 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.662053 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.662082 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.662103 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:07Z","lastTransitionTime":"2025-11-29T00:01:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.764396 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.764442 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.764453 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.764472 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.764484 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:07Z","lastTransitionTime":"2025-11-29T00:01:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.867601 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.867694 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.867712 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.867736 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.867750 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:07Z","lastTransitionTime":"2025-11-29T00:01:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.969843 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.969950 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.969969 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.969995 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:07 crc kubenswrapper[4931]: I1129 00:01:07.970012 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:07Z","lastTransitionTime":"2025-11-29T00:01:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.072976 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.074091 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.074307 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.074502 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.074641 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:08Z","lastTransitionTime":"2025-11-29T00:01:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.178366 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.178626 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.178775 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.178925 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.179028 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:08Z","lastTransitionTime":"2025-11-29T00:01:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.211867 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:08 crc kubenswrapper[4931]: E1129 00:01:08.212068 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.282085 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.282365 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.282472 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.282594 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.282675 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:08Z","lastTransitionTime":"2025-11-29T00:01:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.386267 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.386650 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.386765 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.387102 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.387318 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:08Z","lastTransitionTime":"2025-11-29T00:01:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.490587 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.490639 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.490651 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.490675 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.490690 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:08Z","lastTransitionTime":"2025-11-29T00:01:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.593692 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.593741 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.593760 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.593784 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.593799 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:08Z","lastTransitionTime":"2025-11-29T00:01:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.696957 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.697008 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.697024 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.697044 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.697057 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:08Z","lastTransitionTime":"2025-11-29T00:01:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.799877 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.799958 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.799980 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.800006 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.800025 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:08Z","lastTransitionTime":"2025-11-29T00:01:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.903064 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.903122 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.903144 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.903174 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:08 crc kubenswrapper[4931]: I1129 00:01:08.903192 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:08Z","lastTransitionTime":"2025-11-29T00:01:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.006323 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.006385 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.006403 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.006430 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.006451 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:09Z","lastTransitionTime":"2025-11-29T00:01:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.109906 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.109980 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.110005 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.110040 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.110070 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:09Z","lastTransitionTime":"2025-11-29T00:01:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.211485 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:09 crc kubenswrapper[4931]: E1129 00:01:09.211654 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.211718 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.212076 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:09 crc kubenswrapper[4931]: E1129 00:01:09.212344 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.212589 4931 scope.go:117] "RemoveContainer" containerID="32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232" Nov 29 00:01:09 crc kubenswrapper[4931]: E1129 00:01:09.212750 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.213165 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.213208 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.213226 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.213252 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.213310 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:09Z","lastTransitionTime":"2025-11-29T00:01:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.316009 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.316075 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.316093 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.316121 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.316141 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:09Z","lastTransitionTime":"2025-11-29T00:01:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.422836 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.422880 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.422891 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.423402 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.423425 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:09Z","lastTransitionTime":"2025-11-29T00:01:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.525677 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.525757 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.525781 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.525845 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.525874 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:09Z","lastTransitionTime":"2025-11-29T00:01:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.628704 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.628745 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.628755 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.628772 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.628781 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:09Z","lastTransitionTime":"2025-11-29T00:01:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.731575 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.731631 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.731652 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.731676 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.731694 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:09Z","lastTransitionTime":"2025-11-29T00:01:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.773302 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/2.log" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.776158 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c"} Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.776775 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.788547 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7111f8fe-6cd1-4dc5-99fb-2ce20e485267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53bafde2aa1cbcfcdb36a53eabbf234b09505c13a764ae7ef41659856767dfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.801262 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.812645 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.821389 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.833079 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00e8b190-9f98-45be-9254-2961ad8064a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c38778155a771c99ed4c71c12be76871dc3299129f77a1551fc19f8482e4065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9dc7cc6f5295935a3444f30c164cc63f6b2639a70fd4f45098f28ae9d0f98d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b89687792ce16c3406e8500eb2bfa867fcdc1db444a0623caca0b5033229f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.837121 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.837168 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.837183 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.837220 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.837237 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:09Z","lastTransitionTime":"2025-11-29T00:01:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.850127 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.872413 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.886403 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.899140 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.922932 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.939847 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.939899 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.939929 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.939958 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.939977 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:09Z","lastTransitionTime":"2025-11-29T00:01:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.942103 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:01:01Z\\\",\\\"message\\\":\\\"2025-11-29T00:00:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d\\\\n2025-11-29T00:00:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d to /host/opt/cni/bin/\\\\n2025-11-29T00:00:16Z [verbose] multus-daemon started\\\\n2025-11-29T00:00:16Z [verbose] Readiness Indicator file check\\\\n2025-11-29T00:01:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.966459 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:40Z\\\",\\\"message\\\":\\\" 6599 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:40.145441 6599 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:00:40.145452 6599 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:40.145480 6599 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 00:00:40.145520 6599 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:40.145531 6599 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 00:00:40.145539 6599 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 00:00:40.145571 6599 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:40.145630 6599 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:00:40.145668 6599 factory.go:656] Stopping watch factory\\\\nI1129 00:00:40.145692 6599 ovnkube.go:599] Stopped ovnkube\\\\nI1129 00:00:40.145728 6599 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:00:40.145738 6599 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 00:00:40.145747 6599 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 00:00:40.145792 6599 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF1129 00:00:40.145877 6599 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.985953 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:09 crc kubenswrapper[4931]: I1129 00:01:09.995089 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:09Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.011947 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.022510 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.034064 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.042333 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.042371 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.042382 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.042402 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.042416 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:10Z","lastTransitionTime":"2025-11-29T00:01:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.043192 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.054679 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.144889 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.144918 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.144926 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.144943 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.144953 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:10Z","lastTransitionTime":"2025-11-29T00:01:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.211356 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:10 crc kubenswrapper[4931]: E1129 00:01:10.211494 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.247986 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.248036 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.248057 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.248082 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.248098 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:10Z","lastTransitionTime":"2025-11-29T00:01:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.351494 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.351542 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.351554 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.351573 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.351586 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:10Z","lastTransitionTime":"2025-11-29T00:01:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.454531 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.454577 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.454589 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.454608 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.454622 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:10Z","lastTransitionTime":"2025-11-29T00:01:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.557693 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.557749 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.557762 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.557782 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.557797 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:10Z","lastTransitionTime":"2025-11-29T00:01:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.660777 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.660856 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.660872 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.660896 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.660912 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:10Z","lastTransitionTime":"2025-11-29T00:01:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.763717 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.763790 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.764098 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.764152 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.764174 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:10Z","lastTransitionTime":"2025-11-29T00:01:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.781667 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/3.log" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.782482 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/2.log" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.785802 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" exitCode=1 Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.785893 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c"} Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.785961 4931 scope.go:117] "RemoveContainer" containerID="32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.786601 4931 scope.go:117] "RemoveContainer" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" Nov 29 00:01:10 crc kubenswrapper[4931]: E1129 00:01:10.786898 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.804652 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.819441 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.837745 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.850193 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.871123 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.871153 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.871237 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.871262 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.871487 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.871511 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:10Z","lastTransitionTime":"2025-11-29T00:01:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.889837 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7111f8fe-6cd1-4dc5-99fb-2ce20e485267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53bafde2aa1cbcfcdb36a53eabbf234b09505c13a764ae7ef41659856767dfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.907657 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.923353 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.935672 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.952110 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00e8b190-9f98-45be-9254-2961ad8064a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c38778155a771c99ed4c71c12be76871dc3299129f77a1551fc19f8482e4065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9dc7cc6f5295935a3444f30c164cc63f6b2639a70fd4f45098f28ae9d0f98d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b89687792ce16c3406e8500eb2bfa867fcdc1db444a0623caca0b5033229f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.966782 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.976148 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.976189 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.976199 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.976217 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.976226 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:10Z","lastTransitionTime":"2025-11-29T00:01:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.983534 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:10 crc kubenswrapper[4931]: I1129 00:01:10.999703 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:10Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.014740 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.024210 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.024320 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.024356 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.024436 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.024468 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.024540 4931 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.024603 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:02:15.024582149 +0000 UTC m=+148.186475371 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.024774 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:15.024765272 +0000 UTC m=+148.186658504 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.024882 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.024898 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.024912 4931 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.024938 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-29 00:02:15.024931455 +0000 UTC m=+148.186824687 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.025048 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.025059 4931 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.025067 4931 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.025086 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-29 00:02:15.025080598 +0000 UTC m=+148.186973830 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.025186 4931 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.025209 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-29 00:02:15.02520336 +0000 UTC m=+148.187096592 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.030484 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.046926 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:01:01Z\\\",\\\"message\\\":\\\"2025-11-29T00:00:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d\\\\n2025-11-29T00:00:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d to /host/opt/cni/bin/\\\\n2025-11-29T00:00:16Z [verbose] multus-daemon started\\\\n2025-11-29T00:00:16Z [verbose] Readiness Indicator file check\\\\n2025-11-29T00:01:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.075440 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32719c6a062cd0f2ceb989d72bd207a1862d6a5f72fb7559a7c1481a5852f232\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:00:40Z\\\",\\\"message\\\":\\\" 6599 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:00:40.145441 6599 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:00:40.145452 6599 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:00:40.145480 6599 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1129 00:00:40.145520 6599 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:00:40.145531 6599 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1129 00:00:40.145539 6599 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1129 00:00:40.145571 6599 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1129 00:00:40.145630 6599 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:00:40.145668 6599 factory.go:656] Stopping watch factory\\\\nI1129 00:00:40.145692 6599 ovnkube.go:599] Stopped ovnkube\\\\nI1129 00:00:40.145728 6599 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:00:40.145738 6599 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1129 00:00:40.145747 6599 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1129 00:00:40.145792 6599 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF1129 00:00:40.145877 6599 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:01:10Z\\\",\\\"message\\\":\\\"ler 4 for removal\\\\nI1129 00:01:10.380976 6987 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:01:10.380983 6987 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1129 00:01:10.381017 6987 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:01:10.381027 6987 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:01:10.381035 6987 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:01:10.381044 6987 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 00:01:10.381052 6987 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:01:10.381071 6987 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:01:10.381052 6987 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1129 00:01:10.381264 6987 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1129 00:01:10.381353 6987 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1129 00:01:10.382373 6987 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.079177 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.079266 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.079503 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.079536 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.079547 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:11Z","lastTransitionTime":"2025-11-29T00:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.100543 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.113540 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.182909 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.182966 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.182983 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.183008 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.183026 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:11Z","lastTransitionTime":"2025-11-29T00:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.211726 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.211801 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.211724 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.211921 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.212083 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.212187 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.285653 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.285713 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.285735 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.285769 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.285792 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:11Z","lastTransitionTime":"2025-11-29T00:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.388332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.388395 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.388409 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.388430 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.388762 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:11Z","lastTransitionTime":"2025-11-29T00:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.491655 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.491704 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.491717 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.491738 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.491753 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:11Z","lastTransitionTime":"2025-11-29T00:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.594049 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.594096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.594114 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.594138 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.594156 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:11Z","lastTransitionTime":"2025-11-29T00:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.696321 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.696375 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.696392 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.696414 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.696431 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:11Z","lastTransitionTime":"2025-11-29T00:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.790632 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/3.log" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.794361 4931 scope.go:117] "RemoveContainer" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" Nov 29 00:01:11 crc kubenswrapper[4931]: E1129 00:01:11.794569 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.798486 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.798546 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.798558 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.798578 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.798592 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:11Z","lastTransitionTime":"2025-11-29T00:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.808654 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.819225 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.832400 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00e8b190-9f98-45be-9254-2961ad8064a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c38778155a771c99ed4c71c12be76871dc3299129f77a1551fc19f8482e4065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9dc7cc6f5295935a3444f30c164cc63f6b2639a70fd4f45098f28ae9d0f98d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b89687792ce16c3406e8500eb2bfa867fcdc1db444a0623caca0b5033229f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.843616 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7111f8fe-6cd1-4dc5-99fb-2ce20e485267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53bafde2aa1cbcfcdb36a53eabbf234b09505c13a764ae7ef41659856767dfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.858863 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.872152 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.884413 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.902345 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.902415 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.902431 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.902467 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.902483 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:11Z","lastTransitionTime":"2025-11-29T00:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.904651 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.935396 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:01:01Z\\\",\\\"message\\\":\\\"2025-11-29T00:00:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d\\\\n2025-11-29T00:00:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d to /host/opt/cni/bin/\\\\n2025-11-29T00:00:16Z [verbose] multus-daemon started\\\\n2025-11-29T00:00:16Z [verbose] Readiness Indicator file check\\\\n2025-11-29T00:01:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:11 crc kubenswrapper[4931]: I1129 00:01:11.983669 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:01:10Z\\\",\\\"message\\\":\\\"ler 4 for removal\\\\nI1129 00:01:10.380976 6987 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:01:10.380983 6987 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1129 00:01:10.381017 6987 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:01:10.381027 6987 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:01:10.381035 6987 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:01:10.381044 6987 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 00:01:10.381052 6987 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:01:10.381071 6987 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:01:10.381052 6987 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1129 00:01:10.381264 6987 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1129 00:01:10.381353 6987 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1129 00:01:10.382373 6987 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:01:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:11Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.006510 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.006571 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.006580 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.006595 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.006606 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.032738 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.052843 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.069071 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.085597 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.102893 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.109207 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.109264 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.109278 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.109296 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.109311 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.119615 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.135844 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.152408 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.164947 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.211423 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:12 crc kubenswrapper[4931]: E1129 00:01:12.211636 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.212495 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.212538 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.212557 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.212579 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.212595 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.316202 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.316271 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.316293 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.316316 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.316331 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.418758 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.418838 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.418852 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.418873 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.418887 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.521305 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.521367 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.521386 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.521410 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.521424 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.540234 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.540277 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.540292 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.540312 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.540326 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: E1129 00:01:12.554055 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.560597 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.560693 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.560720 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.560752 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.560778 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: E1129 00:01:12.583048 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.592497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.592540 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.592552 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.592576 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.592587 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: E1129 00:01:12.609789 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.614859 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.614923 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.614941 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.614968 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.614985 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: E1129 00:01:12.631339 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.636953 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.637000 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.637012 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.637033 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.637050 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: E1129 00:01:12.652942 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:12Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:12 crc kubenswrapper[4931]: E1129 00:01:12.653140 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.655895 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.656011 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.656033 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.656082 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.656116 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.758303 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.758377 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.758405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.758441 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.758467 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.861065 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.861123 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.861148 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.861179 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.861201 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.964032 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.964113 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.964141 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.964177 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:12 crc kubenswrapper[4931]: I1129 00:01:12.964200 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:12Z","lastTransitionTime":"2025-11-29T00:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.068144 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.068198 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.068213 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.068236 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.068254 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:13Z","lastTransitionTime":"2025-11-29T00:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.170981 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.171028 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.171046 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.171069 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.171086 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:13Z","lastTransitionTime":"2025-11-29T00:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.211906 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.212089 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.212263 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:13 crc kubenswrapper[4931]: E1129 00:01:13.212280 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:13 crc kubenswrapper[4931]: E1129 00:01:13.212401 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:13 crc kubenswrapper[4931]: E1129 00:01:13.212536 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.274724 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.274765 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.274776 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.274795 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.274821 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:13Z","lastTransitionTime":"2025-11-29T00:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.378008 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.378073 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.378086 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.378102 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.378113 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:13Z","lastTransitionTime":"2025-11-29T00:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.481318 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.481386 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.481405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.481430 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.481450 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:13Z","lastTransitionTime":"2025-11-29T00:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.584627 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.584688 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.584702 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.584726 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.584742 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:13Z","lastTransitionTime":"2025-11-29T00:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.688024 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.688090 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.688108 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.688136 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.688158 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:13Z","lastTransitionTime":"2025-11-29T00:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.790912 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.790965 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.790979 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.791000 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.791015 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:13Z","lastTransitionTime":"2025-11-29T00:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.894685 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.894746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.894757 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.894778 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.894799 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:13Z","lastTransitionTime":"2025-11-29T00:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.998136 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.998222 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.998248 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.998281 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:13 crc kubenswrapper[4931]: I1129 00:01:13.998303 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:13Z","lastTransitionTime":"2025-11-29T00:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.100710 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.100836 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.100864 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.100901 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.100929 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:14Z","lastTransitionTime":"2025-11-29T00:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.204221 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.204289 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.204312 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.204342 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.204365 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:14Z","lastTransitionTime":"2025-11-29T00:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.211751 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:14 crc kubenswrapper[4931]: E1129 00:01:14.212006 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.308061 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.308156 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.308243 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.308275 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.308374 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:14Z","lastTransitionTime":"2025-11-29T00:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.411161 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.411209 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.411223 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.411242 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.411255 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:14Z","lastTransitionTime":"2025-11-29T00:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.513739 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.513787 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.513801 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.513844 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.513858 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:14Z","lastTransitionTime":"2025-11-29T00:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.617072 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.617110 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.617118 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.617132 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.617140 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:14Z","lastTransitionTime":"2025-11-29T00:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.720043 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.720081 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.720089 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.720101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.720110 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:14Z","lastTransitionTime":"2025-11-29T00:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.822315 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.822345 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.822354 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.822366 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.822375 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:14Z","lastTransitionTime":"2025-11-29T00:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.924663 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.924709 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.924723 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.924738 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:14 crc kubenswrapper[4931]: I1129 00:01:14.924749 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:14Z","lastTransitionTime":"2025-11-29T00:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.030750 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.030999 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.031026 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.031044 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.031055 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:15Z","lastTransitionTime":"2025-11-29T00:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.133535 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.133596 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.133619 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.133645 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.133667 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:15Z","lastTransitionTime":"2025-11-29T00:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.212086 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:15 crc kubenswrapper[4931]: E1129 00:01:15.212246 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.212270 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.212108 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:15 crc kubenswrapper[4931]: E1129 00:01:15.212445 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:15 crc kubenswrapper[4931]: E1129 00:01:15.212503 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.237019 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.237081 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.237100 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.237125 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.237157 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:15Z","lastTransitionTime":"2025-11-29T00:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.339987 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.340050 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.340062 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.340079 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.340095 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:15Z","lastTransitionTime":"2025-11-29T00:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.442745 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.442833 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.442853 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.442878 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.442898 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:15Z","lastTransitionTime":"2025-11-29T00:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.545985 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.546113 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.546133 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.546153 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.546172 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:15Z","lastTransitionTime":"2025-11-29T00:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.648690 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.648758 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.648773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.649167 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.649203 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:15Z","lastTransitionTime":"2025-11-29T00:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.751839 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.751907 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.751922 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.751941 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.751953 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:15Z","lastTransitionTime":"2025-11-29T00:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.853995 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.854088 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.854105 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.854125 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.854140 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:15Z","lastTransitionTime":"2025-11-29T00:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.956248 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.956313 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.956332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.956359 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:15 crc kubenswrapper[4931]: I1129 00:01:15.956376 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:15Z","lastTransitionTime":"2025-11-29T00:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.059156 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.059207 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.059219 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.059237 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.059249 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:16Z","lastTransitionTime":"2025-11-29T00:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.162630 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.162667 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.162676 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.162711 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.162721 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:16Z","lastTransitionTime":"2025-11-29T00:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.211650 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:16 crc kubenswrapper[4931]: E1129 00:01:16.211871 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.265761 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.265834 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.265845 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.266011 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.266026 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:16Z","lastTransitionTime":"2025-11-29T00:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.367954 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.368020 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.368037 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.368062 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.368079 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:16Z","lastTransitionTime":"2025-11-29T00:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.471254 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.471289 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.471298 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.471313 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.471322 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:16Z","lastTransitionTime":"2025-11-29T00:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.573738 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.573870 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.573898 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.573928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.573962 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:16Z","lastTransitionTime":"2025-11-29T00:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.676558 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.676607 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.676619 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.676638 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.676650 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:16Z","lastTransitionTime":"2025-11-29T00:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.780050 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.780100 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.780116 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.780134 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.780147 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:16Z","lastTransitionTime":"2025-11-29T00:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.882780 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.883197 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.883217 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.883240 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.883260 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:16Z","lastTransitionTime":"2025-11-29T00:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.986481 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.986543 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.986561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.987012 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:16 crc kubenswrapper[4931]: I1129 00:01:16.987061 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:16Z","lastTransitionTime":"2025-11-29T00:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.089303 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.089339 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.089348 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.089362 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.089373 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:17Z","lastTransitionTime":"2025-11-29T00:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.191320 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.191346 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.191354 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.191380 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.191389 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:17Z","lastTransitionTime":"2025-11-29T00:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.212012 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.212069 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:17 crc kubenswrapper[4931]: E1129 00:01:17.212267 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.212343 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:17 crc kubenswrapper[4931]: E1129 00:01:17.212560 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:17 crc kubenswrapper[4931]: E1129 00:01:17.212695 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.225336 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7111f8fe-6cd1-4dc5-99fb-2ce20e485267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53bafde2aa1cbcfcdb36a53eabbf234b09505c13a764ae7ef41659856767dfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.240991 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.251656 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.262431 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.278932 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00e8b190-9f98-45be-9254-2961ad8064a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c38778155a771c99ed4c71c12be76871dc3299129f77a1551fc19f8482e4065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9dc7cc6f5295935a3444f30c164cc63f6b2639a70fd4f45098f28ae9d0f98d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b89687792ce16c3406e8500eb2bfa867fcdc1db444a0623caca0b5033229f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.295019 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.295066 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.295079 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.295096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.295139 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:17Z","lastTransitionTime":"2025-11-29T00:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.295643 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.311465 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.323996 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.337191 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.351223 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.367332 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:01:01Z\\\",\\\"message\\\":\\\"2025-11-29T00:00:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d\\\\n2025-11-29T00:00:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d to /host/opt/cni/bin/\\\\n2025-11-29T00:00:16Z [verbose] multus-daemon started\\\\n2025-11-29T00:00:16Z [verbose] Readiness Indicator file check\\\\n2025-11-29T00:01:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.394178 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:01:10Z\\\",\\\"message\\\":\\\"ler 4 for removal\\\\nI1129 00:01:10.380976 6987 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:01:10.380983 6987 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1129 00:01:10.381017 6987 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:01:10.381027 6987 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:01:10.381035 6987 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:01:10.381044 6987 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 00:01:10.381052 6987 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:01:10.381071 6987 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:01:10.381052 6987 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1129 00:01:10.381264 6987 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1129 00:01:10.381353 6987 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1129 00:01:10.382373 6987 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:01:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.397873 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.397917 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.397928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.397942 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.397951 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:17Z","lastTransitionTime":"2025-11-29T00:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.423449 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.437628 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.449130 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.460034 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.472289 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.484351 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.497314 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:17Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.500115 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.500161 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.500173 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.500191 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.500203 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:17Z","lastTransitionTime":"2025-11-29T00:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.602396 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.602645 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.602663 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.602684 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.602698 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:17Z","lastTransitionTime":"2025-11-29T00:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.705300 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.705339 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.705350 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.705365 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.705374 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:17Z","lastTransitionTime":"2025-11-29T00:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.807845 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.807900 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.807925 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.807953 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.807973 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:17Z","lastTransitionTime":"2025-11-29T00:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.910694 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.910752 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.910768 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.910791 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:17 crc kubenswrapper[4931]: I1129 00:01:17.910833 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:17Z","lastTransitionTime":"2025-11-29T00:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.013748 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.013797 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.013845 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.013867 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.013882 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:18Z","lastTransitionTime":"2025-11-29T00:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.116473 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.116514 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.116526 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.116543 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.116555 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:18Z","lastTransitionTime":"2025-11-29T00:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.212000 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:18 crc kubenswrapper[4931]: E1129 00:01:18.212195 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.218488 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.218546 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.218563 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.218583 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.218595 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:18Z","lastTransitionTime":"2025-11-29T00:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.321283 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.321337 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.321353 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.321374 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.321388 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:18Z","lastTransitionTime":"2025-11-29T00:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.427022 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.427083 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.427109 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.427133 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.427151 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:18Z","lastTransitionTime":"2025-11-29T00:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.531373 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.531466 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.531491 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.531523 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.531546 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:18Z","lastTransitionTime":"2025-11-29T00:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.634540 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.634593 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.634606 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.634623 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.634637 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:18Z","lastTransitionTime":"2025-11-29T00:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.738208 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.738387 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.738412 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.738437 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.738500 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:18Z","lastTransitionTime":"2025-11-29T00:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.841878 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.841925 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.841969 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.841993 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.842005 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:18Z","lastTransitionTime":"2025-11-29T00:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.945678 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.945733 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.945752 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.945777 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:18 crc kubenswrapper[4931]: I1129 00:01:18.945795 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:18Z","lastTransitionTime":"2025-11-29T00:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.048880 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.048953 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.048977 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.049009 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.049033 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:19Z","lastTransitionTime":"2025-11-29T00:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.151500 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.151574 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.151596 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.151622 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.151643 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:19Z","lastTransitionTime":"2025-11-29T00:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.211372 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.211418 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.211452 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:19 crc kubenswrapper[4931]: E1129 00:01:19.211551 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:19 crc kubenswrapper[4931]: E1129 00:01:19.211598 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:19 crc kubenswrapper[4931]: E1129 00:01:19.211673 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.254312 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.254384 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.254411 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.254444 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.254464 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:19Z","lastTransitionTime":"2025-11-29T00:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.357327 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.357796 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.357909 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.357998 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.358095 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:19Z","lastTransitionTime":"2025-11-29T00:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.460445 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.460783 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.461063 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.461233 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.461365 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:19Z","lastTransitionTime":"2025-11-29T00:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.564387 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.564764 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.565042 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.565268 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.565468 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:19Z","lastTransitionTime":"2025-11-29T00:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.668302 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.668358 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.668368 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.668384 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.668412 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:19Z","lastTransitionTime":"2025-11-29T00:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.770627 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.770933 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.770955 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.770969 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.770980 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:19Z","lastTransitionTime":"2025-11-29T00:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.873122 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.873167 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.873184 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.873206 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.873221 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:19Z","lastTransitionTime":"2025-11-29T00:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.976706 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.977047 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.977253 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.977401 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:19 crc kubenswrapper[4931]: I1129 00:01:19.977566 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:19Z","lastTransitionTime":"2025-11-29T00:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.080880 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.081328 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.081457 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.081553 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.081674 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:20Z","lastTransitionTime":"2025-11-29T00:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.185507 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.185571 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.185589 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.185614 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.185631 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:20Z","lastTransitionTime":"2025-11-29T00:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.212173 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:20 crc kubenswrapper[4931]: E1129 00:01:20.212292 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.289757 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.289861 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.289883 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.289910 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.289934 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:20Z","lastTransitionTime":"2025-11-29T00:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.393511 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.393547 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.393558 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.393576 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.393588 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:20Z","lastTransitionTime":"2025-11-29T00:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.496765 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.496852 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.496868 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.496893 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.496910 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:20Z","lastTransitionTime":"2025-11-29T00:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.600125 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.600176 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.600192 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.600212 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.600225 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:20Z","lastTransitionTime":"2025-11-29T00:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.703840 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.703894 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.703908 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.703934 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.703951 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:20Z","lastTransitionTime":"2025-11-29T00:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.807422 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.807479 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.807496 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.807515 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.807528 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:20Z","lastTransitionTime":"2025-11-29T00:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.911071 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.911137 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.911165 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.911192 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:20 crc kubenswrapper[4931]: I1129 00:01:20.911212 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:20Z","lastTransitionTime":"2025-11-29T00:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.015648 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.015970 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.015997 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.016023 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.016039 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:21Z","lastTransitionTime":"2025-11-29T00:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.120446 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.121020 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.121328 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.121536 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.121725 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:21Z","lastTransitionTime":"2025-11-29T00:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.212906 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:21 crc kubenswrapper[4931]: E1129 00:01:21.215347 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.213953 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:21 crc kubenswrapper[4931]: E1129 00:01:21.215604 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.213787 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:21 crc kubenswrapper[4931]: E1129 00:01:21.215728 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.226868 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.226928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.226949 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.226976 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.226998 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:21Z","lastTransitionTime":"2025-11-29T00:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.330719 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.330800 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.330846 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.330877 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.330896 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:21Z","lastTransitionTime":"2025-11-29T00:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.434557 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.436058 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.436149 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.436177 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.436192 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:21Z","lastTransitionTime":"2025-11-29T00:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.540519 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.540601 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.540620 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.540646 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.540666 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:21Z","lastTransitionTime":"2025-11-29T00:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.643391 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.643433 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.643447 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.643465 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.643478 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:21Z","lastTransitionTime":"2025-11-29T00:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.746171 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.746206 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.746217 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.746232 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.746244 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:21Z","lastTransitionTime":"2025-11-29T00:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.849112 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.849670 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.849802 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.850006 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.850087 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:21Z","lastTransitionTime":"2025-11-29T00:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.953158 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.953864 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.954010 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.954133 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:21 crc kubenswrapper[4931]: I1129 00:01:21.954247 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:21Z","lastTransitionTime":"2025-11-29T00:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.057680 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.057727 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.057742 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.057760 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.057771 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:22Z","lastTransitionTime":"2025-11-29T00:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.161680 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.161745 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.161760 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.161789 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.161832 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:22Z","lastTransitionTime":"2025-11-29T00:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.213217 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:22 crc kubenswrapper[4931]: E1129 00:01:22.213958 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.214435 4931 scope.go:117] "RemoveContainer" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" Nov 29 00:01:22 crc kubenswrapper[4931]: E1129 00:01:22.214722 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.264667 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.264715 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.264752 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.264771 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.264784 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:22Z","lastTransitionTime":"2025-11-29T00:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.367532 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.367596 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.367608 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.367635 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.367651 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:22Z","lastTransitionTime":"2025-11-29T00:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.470707 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.470756 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.470771 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.470794 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.470829 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:22Z","lastTransitionTime":"2025-11-29T00:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.573271 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.574083 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.574101 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.574116 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.574125 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:22Z","lastTransitionTime":"2025-11-29T00:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.676625 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.676696 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.676720 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.676749 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.676771 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:22Z","lastTransitionTime":"2025-11-29T00:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.779386 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.779440 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.779457 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.779476 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.779492 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:22Z","lastTransitionTime":"2025-11-29T00:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.883044 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.883108 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.883127 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.883149 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.883166 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:22Z","lastTransitionTime":"2025-11-29T00:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.985772 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.985845 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.985854 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.985868 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:22 crc kubenswrapper[4931]: I1129 00:01:22.985877 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:22Z","lastTransitionTime":"2025-11-29T00:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.021496 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.021539 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.021556 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.021577 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.021595 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: E1129 00:01:23.041609 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.047096 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.047164 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.047188 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.047215 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.047233 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: E1129 00:01:23.068160 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.073099 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.073176 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.073197 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.073229 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.073250 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: E1129 00:01:23.092944 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.097972 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.098025 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.098040 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.098061 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.098075 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: E1129 00:01:23.115062 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.122859 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.122928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.122953 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.122984 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.123008 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: E1129 00:01:23.142518 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fc84002b-470f-4b68-bb6c-c1b6bbb7e873\\\",\\\"systemUUID\\\":\\\"dfbef0e2-70ff-4485-b415-c8232fbdbdee\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:23Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:23 crc kubenswrapper[4931]: E1129 00:01:23.142711 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.144440 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.144487 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.144504 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.144523 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.144538 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.211757 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.211909 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:23 crc kubenswrapper[4931]: E1129 00:01:23.211951 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:23 crc kubenswrapper[4931]: E1129 00:01:23.212127 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.212278 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:23 crc kubenswrapper[4931]: E1129 00:01:23.212349 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.246991 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.247057 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.247084 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.247112 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.247135 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.349332 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.349382 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.349391 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.349407 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.349416 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.452383 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.452425 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.452466 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.452482 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.452494 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.555251 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.555322 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.555350 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.555380 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.555401 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.657985 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.658043 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.658059 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.658078 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.658092 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.761047 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.761095 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.761105 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.761122 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.761134 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.863624 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.863704 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.863728 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.863760 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.863781 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.966660 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.966713 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.966725 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.966747 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:23 crc kubenswrapper[4931]: I1129 00:01:23.966761 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:23Z","lastTransitionTime":"2025-11-29T00:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.069687 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.069753 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.069770 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.069792 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.069835 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:24Z","lastTransitionTime":"2025-11-29T00:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.173509 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.173587 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.173614 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.173647 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.173672 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:24Z","lastTransitionTime":"2025-11-29T00:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.211566 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:24 crc kubenswrapper[4931]: E1129 00:01:24.211779 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.275533 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.275574 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.275587 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.275605 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.275614 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:24Z","lastTransitionTime":"2025-11-29T00:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.377746 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.377830 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.377845 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.377862 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.377872 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:24Z","lastTransitionTime":"2025-11-29T00:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.480338 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.480375 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.480384 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.480397 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.480406 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:24Z","lastTransitionTime":"2025-11-29T00:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.582252 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.582287 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.582297 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.582314 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.582324 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:24Z","lastTransitionTime":"2025-11-29T00:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.684326 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.684364 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.684372 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.684386 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.684394 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:24Z","lastTransitionTime":"2025-11-29T00:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.786691 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.786741 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.786756 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.786777 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.786794 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:24Z","lastTransitionTime":"2025-11-29T00:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.889089 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.889134 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.889145 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.889161 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.889169 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:24Z","lastTransitionTime":"2025-11-29T00:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.992100 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.992187 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.992212 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.992244 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:24 crc kubenswrapper[4931]: I1129 00:01:24.992269 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:24Z","lastTransitionTime":"2025-11-29T00:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.095363 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.095405 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.095415 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.095431 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.095457 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:25Z","lastTransitionTime":"2025-11-29T00:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.199085 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.199921 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.199966 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.200054 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.200079 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:25Z","lastTransitionTime":"2025-11-29T00:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.213075 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.213139 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:25 crc kubenswrapper[4931]: E1129 00:01:25.213226 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.213077 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:25 crc kubenswrapper[4931]: E1129 00:01:25.213457 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:25 crc kubenswrapper[4931]: E1129 00:01:25.213985 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.303516 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.303574 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.303585 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.303603 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.303615 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:25Z","lastTransitionTime":"2025-11-29T00:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.406224 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.406265 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.406277 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.406299 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.406311 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:25Z","lastTransitionTime":"2025-11-29T00:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.508497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.508536 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.508548 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.508564 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.508574 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:25Z","lastTransitionTime":"2025-11-29T00:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.610959 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.611026 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.611045 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.611070 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.611088 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:25Z","lastTransitionTime":"2025-11-29T00:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.716908 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.716994 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.717016 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.717046 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.717074 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:25Z","lastTransitionTime":"2025-11-29T00:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.819693 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.819729 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.819738 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.819768 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.819778 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:25Z","lastTransitionTime":"2025-11-29T00:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.921883 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.921924 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.921935 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.921951 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:25 crc kubenswrapper[4931]: I1129 00:01:25.921963 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:25Z","lastTransitionTime":"2025-11-29T00:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.024227 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.024284 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.024302 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.024325 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.024342 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:26Z","lastTransitionTime":"2025-11-29T00:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.128974 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.129041 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.129064 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.129175 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.129202 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:26Z","lastTransitionTime":"2025-11-29T00:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.212053 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:26 crc kubenswrapper[4931]: E1129 00:01:26.212183 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.231772 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.231875 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.231893 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.231944 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.231963 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:26Z","lastTransitionTime":"2025-11-29T00:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.334133 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.334208 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.334220 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.334235 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.334247 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:26Z","lastTransitionTime":"2025-11-29T00:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.436838 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.436892 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.436907 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.436928 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.436970 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:26Z","lastTransitionTime":"2025-11-29T00:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.538857 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.539066 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.539084 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.539103 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.539116 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:26Z","lastTransitionTime":"2025-11-29T00:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.641732 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.641792 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.641803 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.641873 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.641883 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:26Z","lastTransitionTime":"2025-11-29T00:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.744088 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.744134 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.744148 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.744169 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.744184 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:26Z","lastTransitionTime":"2025-11-29T00:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.846908 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.846958 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.846972 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.846992 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.847006 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:26Z","lastTransitionTime":"2025-11-29T00:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.949758 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.949828 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.949839 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.949853 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:26 crc kubenswrapper[4931]: I1129 00:01:26.949863 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:26Z","lastTransitionTime":"2025-11-29T00:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.052772 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.052868 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.052892 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.052922 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.052943 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:27Z","lastTransitionTime":"2025-11-29T00:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.156057 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.156105 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.156116 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.156136 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.156148 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:27Z","lastTransitionTime":"2025-11-29T00:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.211252 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.211345 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:27 crc kubenswrapper[4931]: E1129 00:01:27.212232 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.212629 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:27 crc kubenswrapper[4931]: E1129 00:01:27.213191 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:27 crc kubenswrapper[4931]: E1129 00:01:27.213684 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.229454 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4061e63baebbf88b4d0437230e1b88cae07a1937b128d43af3181e730d33213b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2476149ceabd40652e26ad796d57c0ed25caaf95705e29842727d66df35b3dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.259121 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.259180 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.259192 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.259210 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.259222 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:27Z","lastTransitionTime":"2025-11-29T00:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.260049 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.274672 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b227491-f635-4817-96ed-90b96ab6aab3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79c7a95620e4faba1e3ca945128c11b9c47040299a3a3aeb75c1b4bb0962162a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7hhsp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6xxhj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.290901 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc6c5917-e8b2-45a3-95a6-04ccb7ae8fde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4e0ea63c5b13d4aca69756437359fe6f120b260993c5e269b268ed419ffe1e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5debfbab30a25e84fc7dc3f801fc477e51429dcbc0e142da798ab90d2c86cb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9227a79e63dca7a525e065691e44556a415186f7f7b86dd56bd24445e767412\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625813b8cc2d7a39d86b21e2fbd7ffb2a7bc2f680318f8f678a2f7f852e9d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d2acf734f1c7c72892217b6880b6d64404249e9420782035a1b9cf350a5724\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de748df11694ec76472bcdbfeafcdbe0fce1b8deb8032db35b020e77f358a2f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cb7bac031728e602522b2f7bf62b9521b3c9ec0a5325eb1aafb3eb4666490e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szjx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-n5l7h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.304921 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-22vkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31e6835b-462a-4a4f-9ba9-74531c14d41a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:01:01Z\\\",\\\"message\\\":\\\"2025-11-29T00:00:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d\\\\n2025-11-29T00:00:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a5a332a-2476-43d9-91fb-f4f0aea11c4d to /host/opt/cni/bin/\\\\n2025-11-29T00:00:16Z [verbose] multus-daemon started\\\\n2025-11-29T00:00:16Z [verbose] Readiness Indicator file check\\\\n2025-11-29T00:01:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2xl2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-22vkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.335651 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2fae9705-fac9-4c99-b5f8-280e5c05e450\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-29T00:01:10Z\\\",\\\"message\\\":\\\"ler 4 for removal\\\\nI1129 00:01:10.380976 6987 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1129 00:01:10.380983 6987 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1129 00:01:10.381017 6987 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1129 00:01:10.381027 6987 handler.go:208] Removed *v1.Node event handler 2\\\\nI1129 00:01:10.381035 6987 handler.go:208] Removed *v1.Node event handler 7\\\\nI1129 00:01:10.381044 6987 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1129 00:01:10.381052 6987 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1129 00:01:10.381071 6987 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1129 00:01:10.381052 6987 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1129 00:01:10.381264 6987 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1129 00:01:10.381353 6987 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1129 00:01:10.382373 6987 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-29T00:01:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-29T00:00:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nfsrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4m4l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.358333 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"923bd746-f8aa-41d9-b039-d17e5ec582d6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e6629c3362e85fd38d5b724e84f70e8895abf7868b9a82bb2c58d327474b470\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20e6452cc1dc30f938e75b76325ea58ec551a58e22a459da38ed4bd1954d4219\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f405b46e9054f91cfdbb8cde9112143906717f6c772b5c1b9bd9dc18fd4f487a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd694ce25fb94ff742c079656ef2f473b1842b78602b6ceaa6db7d466a425967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560eb2adf161df2909aad5548868edeea64207cb276c8e226db6b0b39f59209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57bbcdada6960b7d34c010b9000aaa56fe47ce1ec5167c3707a827a0a4ac8b57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f95b960e80dc566cea541d69121460aa43ca5cae6a0e67bd77ac423d7765d087\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a649566f969313bba33e9cd035b06bce9d094d3b6eff2086eb4ff7860e33f4dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.361178 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.361259 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.361277 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.361331 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.361354 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:27Z","lastTransitionTime":"2025-11-29T00:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.375450 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a82fabf-b604-45cb-9526-aba5e6ff2fa0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://380990c08faa146c4ffdfc91e192cdea1f14b54a93f507552c67a774d3676794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8e273142a1299f060f0af57ab1f58a30630bb6d2b6366abb36e415d97bef02e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://670b6db57e4dba38d36659056725a822aa1642c733367cd26fc0eca4c784009a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.384990 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e38ad69-f5fb-415d-ad76-e899aed032a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-848rn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7dwtb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.394557 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5767ea5-e04c-458d-a4bf-ca8941144fcb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfd5d353d62a5a67c57e782c9ff9a6d2be84b9178f97da36bc2fdfbf3e144ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b8147730542b715b931ac5274428cb634fa8b89789d5f5d20c8383d03745c32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bvsm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bs5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.404342 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.415547 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.427674 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbf400a22a85b89b7352eecf890fb36f6de4e00a691175c3a6e66a59dc564a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.438443 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-n2tnq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c2ac4c77-c7aa-40ec-8bd9-e6afa4b2a8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20365f4d761dce82e4f294b78ecebfb3497c3dad9ba8f8058db2d0c4e705a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzcx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-n2tnq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.457135 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-29T00:00:06Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1129 00:00:00.734220 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1129 00:00:00.735740 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-617461892/tls.crt::/tmp/serving-cert-617461892/tls.key\\\\\\\"\\\\nI1129 00:00:06.163183 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1129 00:00:06.166545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1129 00:00:06.166569 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1129 00:00:06.166602 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1129 00:00:06.166610 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1129 00:00:06.177292 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1129 00:00:06.177338 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1129 00:00:06.177435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177458 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1129 00:00:06.177478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1129 00:00:06.177508 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1129 00:00:06.177529 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1129 00:00:06.177548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1129 00:00:06.182190 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.464335 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.464398 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.464421 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.464497 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.464523 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:27Z","lastTransitionTime":"2025-11-29T00:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.473013 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5239e876386c58ad987438bdaed32321d65c98eac47aac077edcae8926e1dc15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.484307 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w99nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f08fd4c-2846-497a-955d-c66ba5b58a9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be9f3561fdd844d40203be91d02f599b9487783e68c49761ebaa4f9bf0b46616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-29T00:00:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjjkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-29T00:00:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w99nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.494685 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00e8b190-9f98-45be-9254-2961ad8064a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-29T00:00:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c38778155a771c99ed4c71c12be76871dc3299129f77a1551fc19f8482e4065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9dc7cc6f5295935a3444f30c164cc63f6b2639a70fd4f45098f28ae9d0f98d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b89687792ce16c3406e8500eb2bfa867fcdc1db444a0623caca0b5033229f33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047a3934b5f9d77365c716e29a4a5606cddc61002014b1cfff2db334ffe7ed52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.503786 4931 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7111f8fe-6cd1-4dc5-99fb-2ce20e485267\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-28T23:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53bafde2aa1cbcfcdb36a53eabbf234b09505c13a764ae7ef41659856767dfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-28T23:59:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a451aa4dd7ff6c9863f8e63c5c69696664b9df3496959cfe39b43ec929c2a264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-28T23:59:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-28T23:59:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-28T23:59:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-29T00:01:27Z is after 2025-08-24T17:21:41Z" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.565979 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.566501 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.566545 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.566561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.566570 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:27Z","lastTransitionTime":"2025-11-29T00:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.670239 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.670307 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.670325 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.670351 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.670372 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:27Z","lastTransitionTime":"2025-11-29T00:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.773360 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.773417 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.773434 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.773456 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.773469 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:27Z","lastTransitionTime":"2025-11-29T00:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.876224 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.876328 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.876351 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.876380 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.876401 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:27Z","lastTransitionTime":"2025-11-29T00:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.978594 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.978648 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.978663 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.978685 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:27 crc kubenswrapper[4931]: I1129 00:01:27.978701 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:27Z","lastTransitionTime":"2025-11-29T00:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.080779 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.080857 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.080875 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.080898 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.080915 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:28Z","lastTransitionTime":"2025-11-29T00:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.183426 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.183469 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.183479 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.183496 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.183511 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:28Z","lastTransitionTime":"2025-11-29T00:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.212145 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:28 crc kubenswrapper[4931]: E1129 00:01:28.212266 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.285744 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.285802 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.285837 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.285857 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.285869 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:28Z","lastTransitionTime":"2025-11-29T00:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.388614 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.388672 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.388684 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.388704 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.388718 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:28Z","lastTransitionTime":"2025-11-29T00:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.491012 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.491088 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.491114 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.491146 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.491170 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:28Z","lastTransitionTime":"2025-11-29T00:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.593493 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.593528 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.593539 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.593554 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.593565 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:28Z","lastTransitionTime":"2025-11-29T00:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.696266 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.696316 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.696336 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.696354 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.696369 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:28Z","lastTransitionTime":"2025-11-29T00:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.798466 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.798507 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.798521 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.798543 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.798559 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:28Z","lastTransitionTime":"2025-11-29T00:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.902001 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.902075 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.902090 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.902110 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:28 crc kubenswrapper[4931]: I1129 00:01:28.902134 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:28Z","lastTransitionTime":"2025-11-29T00:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.005359 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.005420 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.005432 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.005451 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.005463 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:29Z","lastTransitionTime":"2025-11-29T00:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.108463 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.108531 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.108542 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.108561 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.108573 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:29Z","lastTransitionTime":"2025-11-29T00:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.211268 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.211379 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.211412 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.211485 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:29 crc kubenswrapper[4931]: E1129 00:01:29.211514 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.211527 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.211591 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.211688 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.211699 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:29Z","lastTransitionTime":"2025-11-29T00:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:29 crc kubenswrapper[4931]: E1129 00:01:29.211745 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:29 crc kubenswrapper[4931]: E1129 00:01:29.211915 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.314075 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.314128 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.314140 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.314156 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.314166 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:29Z","lastTransitionTime":"2025-11-29T00:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.417218 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.417302 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.417322 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.417347 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.417366 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:29Z","lastTransitionTime":"2025-11-29T00:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.519409 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.519486 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.519499 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.519519 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.519532 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:29Z","lastTransitionTime":"2025-11-29T00:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.622213 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.622255 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.622267 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.622283 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.622297 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:29Z","lastTransitionTime":"2025-11-29T00:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.725630 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.725673 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.725686 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.725705 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.725719 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:29Z","lastTransitionTime":"2025-11-29T00:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.828137 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.828190 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.828202 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.828222 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.828234 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:29Z","lastTransitionTime":"2025-11-29T00:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.931085 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.931137 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.931148 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.931168 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:29 crc kubenswrapper[4931]: I1129 00:01:29.931181 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:29Z","lastTransitionTime":"2025-11-29T00:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.034140 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.034198 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.034214 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.034244 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.034260 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:30Z","lastTransitionTime":"2025-11-29T00:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.136444 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.136511 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.136529 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.136552 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.136571 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:30Z","lastTransitionTime":"2025-11-29T00:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.212296 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:30 crc kubenswrapper[4931]: E1129 00:01:30.212622 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.239083 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.239165 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.239182 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.239201 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.239214 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:30Z","lastTransitionTime":"2025-11-29T00:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.340898 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.340942 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.340956 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.340976 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.340990 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:30Z","lastTransitionTime":"2025-11-29T00:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.443124 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.443160 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.443173 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.443189 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.443201 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:30Z","lastTransitionTime":"2025-11-29T00:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.545070 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.545113 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.545125 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.545140 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.545153 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:30Z","lastTransitionTime":"2025-11-29T00:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.647699 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.647743 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.647755 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.647773 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.647785 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:30Z","lastTransitionTime":"2025-11-29T00:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.749638 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.749680 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.749691 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.749708 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.749725 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:30Z","lastTransitionTime":"2025-11-29T00:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.852221 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.852275 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.852290 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.852311 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.852328 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:30Z","lastTransitionTime":"2025-11-29T00:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.955083 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.955142 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.955157 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.955178 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:30 crc kubenswrapper[4931]: I1129 00:01:30.955194 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:30Z","lastTransitionTime":"2025-11-29T00:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.057924 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.057981 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.057993 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.058012 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.058024 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:31Z","lastTransitionTime":"2025-11-29T00:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.160597 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.160661 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.160682 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.160705 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.160720 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:31Z","lastTransitionTime":"2025-11-29T00:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.211354 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.211396 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.211354 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:31 crc kubenswrapper[4931]: E1129 00:01:31.211561 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:31 crc kubenswrapper[4931]: E1129 00:01:31.211785 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:31 crc kubenswrapper[4931]: E1129 00:01:31.212102 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.263168 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.263210 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.263218 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.263232 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.263242 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:31Z","lastTransitionTime":"2025-11-29T00:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.368668 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.368727 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.368740 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.368756 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.368784 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:31Z","lastTransitionTime":"2025-11-29T00:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.471310 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.471371 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.471382 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.471397 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.471407 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:31Z","lastTransitionTime":"2025-11-29T00:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.573792 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.573844 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.573855 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.573871 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.573882 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:31Z","lastTransitionTime":"2025-11-29T00:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.676382 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.676426 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.676439 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.676457 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.676468 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:31Z","lastTransitionTime":"2025-11-29T00:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.778903 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.778971 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.778995 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.779024 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.779046 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:31Z","lastTransitionTime":"2025-11-29T00:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.881978 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.882038 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.882059 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.882081 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.882097 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:31Z","lastTransitionTime":"2025-11-29T00:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.985154 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.985217 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.985250 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.985274 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:31 crc kubenswrapper[4931]: I1129 00:01:31.985302 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:31Z","lastTransitionTime":"2025-11-29T00:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.088645 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.088712 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.088737 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.088767 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.088788 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:32Z","lastTransitionTime":"2025-11-29T00:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.191644 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.191740 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.191754 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.191772 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.191785 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:32Z","lastTransitionTime":"2025-11-29T00:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.211355 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:32 crc kubenswrapper[4931]: E1129 00:01:32.211691 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.295316 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.295363 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.295372 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.295387 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.295397 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:32Z","lastTransitionTime":"2025-11-29T00:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.398271 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.398337 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.398360 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.398390 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.398411 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:32Z","lastTransitionTime":"2025-11-29T00:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.501414 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.501478 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.501500 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.501530 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.501554 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:32Z","lastTransitionTime":"2025-11-29T00:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.604364 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.604425 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.604440 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.604461 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.604476 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:32Z","lastTransitionTime":"2025-11-29T00:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.707228 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.707280 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.707297 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.707320 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.707335 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:32Z","lastTransitionTime":"2025-11-29T00:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.811216 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.811275 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.811291 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.811317 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.811359 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:32Z","lastTransitionTime":"2025-11-29T00:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.914139 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.914201 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.914223 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.914251 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:32 crc kubenswrapper[4931]: I1129 00:01:32.914272 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:32Z","lastTransitionTime":"2025-11-29T00:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.016604 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.016638 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.016649 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.016665 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.016677 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:33Z","lastTransitionTime":"2025-11-29T00:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.119761 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.119860 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.119879 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.119906 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.119925 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:33Z","lastTransitionTime":"2025-11-29T00:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.160590 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:33 crc kubenswrapper[4931]: E1129 00:01:33.160868 4931 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:01:33 crc kubenswrapper[4931]: E1129 00:01:33.160979 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs podName:5e38ad69-f5fb-415d-ad76-e899aed032a6 nodeName:}" failed. No retries permitted until 2025-11-29 00:02:37.160950326 +0000 UTC m=+170.322843598 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs") pod "network-metrics-daemon-7dwtb" (UID: "5e38ad69-f5fb-415d-ad76-e899aed032a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.212992 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.213068 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.213121 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:33 crc kubenswrapper[4931]: E1129 00:01:33.213284 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:33 crc kubenswrapper[4931]: E1129 00:01:33.213393 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:33 crc kubenswrapper[4931]: E1129 00:01:33.213473 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.222460 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.222524 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.222542 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.222565 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.222584 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:33Z","lastTransitionTime":"2025-11-29T00:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.266846 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.266915 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.266939 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.266973 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.266995 4931 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-29T00:01:33Z","lastTransitionTime":"2025-11-29T00:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.339189 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4"] Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.339856 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.342978 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.345035 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.345460 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.346995 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.358532 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-w99nx" podStartSLOduration=80.358501432 podStartE2EDuration="1m20.358501432s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:01:33.358212447 +0000 UTC m=+106.520105719" watchObservedRunningTime="2025-11-29 00:01:33.358501432 +0000 UTC m=+106.520394704" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.362309 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0872d728-775d-4b0c-928d-82aaafee5de7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.362376 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0872d728-775d-4b0c-928d-82aaafee5de7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.362443 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0872d728-775d-4b0c-928d-82aaafee5de7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.362515 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0872d728-775d-4b0c-928d-82aaafee5de7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.362579 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0872d728-775d-4b0c-928d-82aaafee5de7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.399377 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=50.399344369 podStartE2EDuration="50.399344369s" podCreationTimestamp="2025-11-29 00:00:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:01:33.398946383 +0000 UTC m=+106.560839665" watchObservedRunningTime="2025-11-29 00:01:33.399344369 +0000 UTC m=+106.561237631" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.414608 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=38.414572566 podStartE2EDuration="38.414572566s" podCreationTimestamp="2025-11-29 00:00:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:01:33.414383982 +0000 UTC m=+106.576277224" watchObservedRunningTime="2025-11-29 00:01:33.414572566 +0000 UTC m=+106.576465838" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.438383 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=87.438358595 podStartE2EDuration="1m27.438358595s" podCreationTimestamp="2025-11-29 00:00:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:01:33.43680175 +0000 UTC m=+106.598695082" watchObservedRunningTime="2025-11-29 00:01:33.438358595 +0000 UTC m=+106.600251837" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.463594 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0872d728-775d-4b0c-928d-82aaafee5de7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.463658 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0872d728-775d-4b0c-928d-82aaafee5de7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.463720 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0872d728-775d-4b0c-928d-82aaafee5de7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.463757 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0872d728-775d-4b0c-928d-82aaafee5de7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.463798 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0872d728-775d-4b0c-928d-82aaafee5de7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.463898 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0872d728-775d-4b0c-928d-82aaafee5de7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.464136 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0872d728-775d-4b0c-928d-82aaafee5de7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.465286 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0872d728-775d-4b0c-928d-82aaafee5de7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.469767 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0872d728-775d-4b0c-928d-82aaafee5de7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.493587 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0872d728-775d-4b0c-928d-82aaafee5de7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xl8r4\" (UID: \"0872d728-775d-4b0c-928d-82aaafee5de7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.503616 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-n5l7h" podStartSLOduration=80.503595183 podStartE2EDuration="1m20.503595183s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:01:33.50339071 +0000 UTC m=+106.665283982" watchObservedRunningTime="2025-11-29 00:01:33.503595183 +0000 UTC m=+106.665488415" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.503841 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podStartSLOduration=80.503800117 podStartE2EDuration="1m20.503800117s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:01:33.47665309 +0000 UTC m=+106.638546362" watchObservedRunningTime="2025-11-29 00:01:33.503800117 +0000 UTC m=+106.665693349" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.523778 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-22vkg" podStartSLOduration=80.523753273 podStartE2EDuration="1m20.523753273s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:01:33.523302665 +0000 UTC m=+106.685195907" watchObservedRunningTime="2025-11-29 00:01:33.523753273 +0000 UTC m=+106.685646535" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.586441 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=84.586422908 podStartE2EDuration="1m24.586422908s" podCreationTimestamp="2025-11-29 00:00:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:01:33.585471971 +0000 UTC m=+106.747365243" watchObservedRunningTime="2025-11-29 00:01:33.586422908 +0000 UTC m=+106.748316150" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.607724 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=86.607702836 podStartE2EDuration="1m26.607702836s" podCreationTimestamp="2025-11-29 00:00:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:01:33.606509046 +0000 UTC m=+106.768402298" watchObservedRunningTime="2025-11-29 00:01:33.607702836 +0000 UTC m=+106.769596078" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.664847 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" Nov 29 00:01:33 crc kubenswrapper[4931]: W1129 00:01:33.694292 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0872d728_775d_4b0c_928d_82aaafee5de7.slice/crio-bb5cb0def6fd1c4d0a3f6899f96d4956bddb7e6f5587420370b0b0342f63d91e WatchSource:0}: Error finding container bb5cb0def6fd1c4d0a3f6899f96d4956bddb7e6f5587420370b0b0342f63d91e: Status 404 returned error can't find the container with id bb5cb0def6fd1c4d0a3f6899f96d4956bddb7e6f5587420370b0b0342f63d91e Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.713407 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bs5dt" podStartSLOduration=80.713386145 podStartE2EDuration="1m20.713386145s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:01:33.712662173 +0000 UTC m=+106.874555425" watchObservedRunningTime="2025-11-29 00:01:33.713386145 +0000 UTC m=+106.875279397" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.747513 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-n2tnq" podStartSLOduration=80.747488829 podStartE2EDuration="1m20.747488829s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:01:33.746507012 +0000 UTC m=+106.908400254" watchObservedRunningTime="2025-11-29 00:01:33.747488829 +0000 UTC m=+106.909382081" Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.873053 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" event={"ID":"0872d728-775d-4b0c-928d-82aaafee5de7","Type":"ContainerStarted","Data":"d32b9e0253cf07a91975058432514fc52809c2e55d94bb7ae593109fbc794966"} Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.873130 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" event={"ID":"0872d728-775d-4b0c-928d-82aaafee5de7","Type":"ContainerStarted","Data":"bb5cb0def6fd1c4d0a3f6899f96d4956bddb7e6f5587420370b0b0342f63d91e"} Nov 29 00:01:33 crc kubenswrapper[4931]: I1129 00:01:33.891417 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xl8r4" podStartSLOduration=80.891396051 podStartE2EDuration="1m20.891396051s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:01:33.890998695 +0000 UTC m=+107.052891947" watchObservedRunningTime="2025-11-29 00:01:33.891396051 +0000 UTC m=+107.053289313" Nov 29 00:01:34 crc kubenswrapper[4931]: I1129 00:01:34.212854 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:34 crc kubenswrapper[4931]: E1129 00:01:34.213064 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:35 crc kubenswrapper[4931]: I1129 00:01:35.211402 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:35 crc kubenswrapper[4931]: I1129 00:01:35.211415 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:35 crc kubenswrapper[4931]: I1129 00:01:35.211475 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:35 crc kubenswrapper[4931]: E1129 00:01:35.213121 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:35 crc kubenswrapper[4931]: E1129 00:01:35.213262 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:35 crc kubenswrapper[4931]: E1129 00:01:35.213410 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:36 crc kubenswrapper[4931]: I1129 00:01:36.211455 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:36 crc kubenswrapper[4931]: E1129 00:01:36.211633 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:37 crc kubenswrapper[4931]: I1129 00:01:37.211638 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:37 crc kubenswrapper[4931]: I1129 00:01:37.211687 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:37 crc kubenswrapper[4931]: I1129 00:01:37.214760 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:37 crc kubenswrapper[4931]: E1129 00:01:37.214954 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:37 crc kubenswrapper[4931]: E1129 00:01:37.215082 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:37 crc kubenswrapper[4931]: E1129 00:01:37.215182 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:37 crc kubenswrapper[4931]: I1129 00:01:37.216461 4931 scope.go:117] "RemoveContainer" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" Nov 29 00:01:37 crc kubenswrapper[4931]: E1129 00:01:37.217193 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" Nov 29 00:01:38 crc kubenswrapper[4931]: I1129 00:01:38.211847 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:38 crc kubenswrapper[4931]: E1129 00:01:38.212032 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:39 crc kubenswrapper[4931]: I1129 00:01:39.211317 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:39 crc kubenswrapper[4931]: E1129 00:01:39.211534 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:39 crc kubenswrapper[4931]: I1129 00:01:39.212057 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:39 crc kubenswrapper[4931]: I1129 00:01:39.212169 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:39 crc kubenswrapper[4931]: E1129 00:01:39.212454 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:39 crc kubenswrapper[4931]: E1129 00:01:39.212769 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:40 crc kubenswrapper[4931]: I1129 00:01:40.212123 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:40 crc kubenswrapper[4931]: E1129 00:01:40.212379 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:41 crc kubenswrapper[4931]: I1129 00:01:41.211315 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:41 crc kubenswrapper[4931]: E1129 00:01:41.211515 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:41 crc kubenswrapper[4931]: I1129 00:01:41.211342 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:41 crc kubenswrapper[4931]: I1129 00:01:41.211570 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:41 crc kubenswrapper[4931]: E1129 00:01:41.211668 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:41 crc kubenswrapper[4931]: E1129 00:01:41.212001 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:42 crc kubenswrapper[4931]: I1129 00:01:42.211991 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:42 crc kubenswrapper[4931]: E1129 00:01:42.212151 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:43 crc kubenswrapper[4931]: I1129 00:01:43.211749 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:43 crc kubenswrapper[4931]: I1129 00:01:43.211759 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:43 crc kubenswrapper[4931]: E1129 00:01:43.212041 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:43 crc kubenswrapper[4931]: I1129 00:01:43.212105 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:43 crc kubenswrapper[4931]: E1129 00:01:43.212643 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:43 crc kubenswrapper[4931]: E1129 00:01:43.213058 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:44 crc kubenswrapper[4931]: I1129 00:01:44.211793 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:44 crc kubenswrapper[4931]: E1129 00:01:44.212028 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:45 crc kubenswrapper[4931]: I1129 00:01:45.211248 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:45 crc kubenswrapper[4931]: I1129 00:01:45.211292 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:45 crc kubenswrapper[4931]: I1129 00:01:45.211330 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:45 crc kubenswrapper[4931]: E1129 00:01:45.211361 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:45 crc kubenswrapper[4931]: E1129 00:01:45.211570 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:45 crc kubenswrapper[4931]: E1129 00:01:45.211683 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:46 crc kubenswrapper[4931]: I1129 00:01:46.212238 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:46 crc kubenswrapper[4931]: E1129 00:01:46.212428 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:47 crc kubenswrapper[4931]: E1129 00:01:47.166213 4931 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 29 00:01:47 crc kubenswrapper[4931]: I1129 00:01:47.212928 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:47 crc kubenswrapper[4931]: E1129 00:01:47.213068 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:47 crc kubenswrapper[4931]: I1129 00:01:47.212931 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:47 crc kubenswrapper[4931]: I1129 00:01:47.213146 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:47 crc kubenswrapper[4931]: E1129 00:01:47.213339 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:47 crc kubenswrapper[4931]: E1129 00:01:47.213519 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:47 crc kubenswrapper[4931]: E1129 00:01:47.602047 4931 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 29 00:01:47 crc kubenswrapper[4931]: I1129 00:01:47.935451 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22vkg_31e6835b-462a-4a4f-9ba9-74531c14d41a/kube-multus/1.log" Nov 29 00:01:47 crc kubenswrapper[4931]: I1129 00:01:47.937398 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22vkg_31e6835b-462a-4a4f-9ba9-74531c14d41a/kube-multus/0.log" Nov 29 00:01:47 crc kubenswrapper[4931]: I1129 00:01:47.937541 4931 generic.go:334] "Generic (PLEG): container finished" podID="31e6835b-462a-4a4f-9ba9-74531c14d41a" containerID="7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652" exitCode=1 Nov 29 00:01:47 crc kubenswrapper[4931]: I1129 00:01:47.937595 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22vkg" event={"ID":"31e6835b-462a-4a4f-9ba9-74531c14d41a","Type":"ContainerDied","Data":"7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652"} Nov 29 00:01:47 crc kubenswrapper[4931]: I1129 00:01:47.937658 4931 scope.go:117] "RemoveContainer" containerID="eedff8502a246096b289ca7e5a1a19990fec5753183e5b334b9b3b0510478857" Nov 29 00:01:47 crc kubenswrapper[4931]: I1129 00:01:47.939474 4931 scope.go:117] "RemoveContainer" containerID="7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652" Nov 29 00:01:47 crc kubenswrapper[4931]: E1129 00:01:47.939849 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-22vkg_openshift-multus(31e6835b-462a-4a4f-9ba9-74531c14d41a)\"" pod="openshift-multus/multus-22vkg" podUID="31e6835b-462a-4a4f-9ba9-74531c14d41a" Nov 29 00:01:48 crc kubenswrapper[4931]: I1129 00:01:48.211304 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:48 crc kubenswrapper[4931]: E1129 00:01:48.211518 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:48 crc kubenswrapper[4931]: I1129 00:01:48.943612 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22vkg_31e6835b-462a-4a4f-9ba9-74531c14d41a/kube-multus/1.log" Nov 29 00:01:49 crc kubenswrapper[4931]: I1129 00:01:49.211730 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:49 crc kubenswrapper[4931]: E1129 00:01:49.211954 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:49 crc kubenswrapper[4931]: I1129 00:01:49.212099 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:49 crc kubenswrapper[4931]: I1129 00:01:49.212115 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:49 crc kubenswrapper[4931]: E1129 00:01:49.212463 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:49 crc kubenswrapper[4931]: E1129 00:01:49.212663 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:49 crc kubenswrapper[4931]: I1129 00:01:49.214988 4931 scope.go:117] "RemoveContainer" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" Nov 29 00:01:49 crc kubenswrapper[4931]: E1129 00:01:49.215474 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4m4l4_openshift-ovn-kubernetes(2fae9705-fac9-4c99-b5f8-280e5c05e450)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" Nov 29 00:01:50 crc kubenswrapper[4931]: I1129 00:01:50.211858 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:50 crc kubenswrapper[4931]: E1129 00:01:50.212063 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:51 crc kubenswrapper[4931]: I1129 00:01:51.211838 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:51 crc kubenswrapper[4931]: I1129 00:01:51.211978 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:51 crc kubenswrapper[4931]: E1129 00:01:51.212117 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:51 crc kubenswrapper[4931]: E1129 00:01:51.212288 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:51 crc kubenswrapper[4931]: I1129 00:01:51.212931 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:51 crc kubenswrapper[4931]: E1129 00:01:51.213080 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:52 crc kubenswrapper[4931]: I1129 00:01:52.211289 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:52 crc kubenswrapper[4931]: E1129 00:01:52.211892 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:52 crc kubenswrapper[4931]: E1129 00:01:52.603741 4931 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 29 00:01:53 crc kubenswrapper[4931]: I1129 00:01:53.212311 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:53 crc kubenswrapper[4931]: I1129 00:01:53.212390 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:53 crc kubenswrapper[4931]: E1129 00:01:53.212588 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:53 crc kubenswrapper[4931]: I1129 00:01:53.212674 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:53 crc kubenswrapper[4931]: E1129 00:01:53.212888 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:53 crc kubenswrapper[4931]: E1129 00:01:53.213048 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:54 crc kubenswrapper[4931]: I1129 00:01:54.211530 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:54 crc kubenswrapper[4931]: E1129 00:01:54.211688 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:55 crc kubenswrapper[4931]: I1129 00:01:55.211435 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:55 crc kubenswrapper[4931]: I1129 00:01:55.211549 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:55 crc kubenswrapper[4931]: E1129 00:01:55.211625 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:55 crc kubenswrapper[4931]: E1129 00:01:55.211729 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:55 crc kubenswrapper[4931]: I1129 00:01:55.212073 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:55 crc kubenswrapper[4931]: E1129 00:01:55.212289 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:56 crc kubenswrapper[4931]: I1129 00:01:56.211447 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:56 crc kubenswrapper[4931]: E1129 00:01:56.211640 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:57 crc kubenswrapper[4931]: I1129 00:01:57.211393 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:57 crc kubenswrapper[4931]: E1129 00:01:57.211555 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:01:57 crc kubenswrapper[4931]: I1129 00:01:57.211696 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:57 crc kubenswrapper[4931]: I1129 00:01:57.211777 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:57 crc kubenswrapper[4931]: E1129 00:01:57.213792 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:57 crc kubenswrapper[4931]: E1129 00:01:57.213951 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:57 crc kubenswrapper[4931]: E1129 00:01:57.605221 4931 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 29 00:01:58 crc kubenswrapper[4931]: I1129 00:01:58.211569 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:01:58 crc kubenswrapper[4931]: E1129 00:01:58.211767 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:01:59 crc kubenswrapper[4931]: I1129 00:01:59.211390 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:01:59 crc kubenswrapper[4931]: E1129 00:01:59.211518 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:01:59 crc kubenswrapper[4931]: I1129 00:01:59.211401 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:01:59 crc kubenswrapper[4931]: E1129 00:01:59.211650 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:01:59 crc kubenswrapper[4931]: I1129 00:01:59.211396 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:01:59 crc kubenswrapper[4931]: E1129 00:01:59.211710 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:02:00 crc kubenswrapper[4931]: I1129 00:02:00.211177 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:02:00 crc kubenswrapper[4931]: E1129 00:02:00.211393 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:02:01 crc kubenswrapper[4931]: I1129 00:02:01.211967 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:02:01 crc kubenswrapper[4931]: E1129 00:02:01.212168 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:02:01 crc kubenswrapper[4931]: I1129 00:02:01.211991 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:02:01 crc kubenswrapper[4931]: I1129 00:02:01.212264 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:02:01 crc kubenswrapper[4931]: E1129 00:02:01.212404 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:02:01 crc kubenswrapper[4931]: E1129 00:02:01.212483 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:02:02 crc kubenswrapper[4931]: I1129 00:02:02.212386 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:02:02 crc kubenswrapper[4931]: I1129 00:02:02.212786 4931 scope.go:117] "RemoveContainer" containerID="7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652" Nov 29 00:02:02 crc kubenswrapper[4931]: E1129 00:02:02.213694 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:02:02 crc kubenswrapper[4931]: E1129 00:02:02.606122 4931 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 29 00:02:03 crc kubenswrapper[4931]: I1129 00:02:03.009392 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22vkg_31e6835b-462a-4a4f-9ba9-74531c14d41a/kube-multus/1.log" Nov 29 00:02:03 crc kubenswrapper[4931]: I1129 00:02:03.009800 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22vkg" event={"ID":"31e6835b-462a-4a4f-9ba9-74531c14d41a","Type":"ContainerStarted","Data":"8759f244d50ee7187a9b3f8c435be8a5a62e4a10fb675c1dc1e1a29cac90118b"} Nov 29 00:02:03 crc kubenswrapper[4931]: I1129 00:02:03.212261 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:02:03 crc kubenswrapper[4931]: I1129 00:02:03.212336 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:02:03 crc kubenswrapper[4931]: E1129 00:02:03.212444 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:02:03 crc kubenswrapper[4931]: I1129 00:02:03.212536 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:02:03 crc kubenswrapper[4931]: E1129 00:02:03.213107 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:02:03 crc kubenswrapper[4931]: E1129 00:02:03.213351 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:02:03 crc kubenswrapper[4931]: I1129 00:02:03.213582 4931 scope.go:117] "RemoveContainer" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" Nov 29 00:02:04 crc kubenswrapper[4931]: I1129 00:02:04.017097 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/3.log" Nov 29 00:02:04 crc kubenswrapper[4931]: I1129 00:02:04.020646 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerStarted","Data":"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946"} Nov 29 00:02:04 crc kubenswrapper[4931]: I1129 00:02:04.021241 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:02:04 crc kubenswrapper[4931]: I1129 00:02:04.068403 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podStartSLOduration=111.068372657 podStartE2EDuration="1m51.068372657s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:04.066921765 +0000 UTC m=+137.228815087" watchObservedRunningTime="2025-11-29 00:02:04.068372657 +0000 UTC m=+137.230265919" Nov 29 00:02:04 crc kubenswrapper[4931]: I1129 00:02:04.078317 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7dwtb"] Nov 29 00:02:04 crc kubenswrapper[4931]: I1129 00:02:04.078497 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:02:04 crc kubenswrapper[4931]: E1129 00:02:04.078688 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:02:04 crc kubenswrapper[4931]: I1129 00:02:04.212122 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:02:04 crc kubenswrapper[4931]: E1129 00:02:04.212356 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:02:05 crc kubenswrapper[4931]: I1129 00:02:05.211886 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:02:05 crc kubenswrapper[4931]: E1129 00:02:05.212038 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:02:05 crc kubenswrapper[4931]: I1129 00:02:05.211895 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:02:05 crc kubenswrapper[4931]: I1129 00:02:05.212105 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:02:05 crc kubenswrapper[4931]: E1129 00:02:05.212229 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:02:05 crc kubenswrapper[4931]: E1129 00:02:05.212657 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:02:06 crc kubenswrapper[4931]: I1129 00:02:06.211500 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:02:06 crc kubenswrapper[4931]: E1129 00:02:06.211701 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 29 00:02:07 crc kubenswrapper[4931]: I1129 00:02:07.212288 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:02:07 crc kubenswrapper[4931]: I1129 00:02:07.212365 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:02:07 crc kubenswrapper[4931]: I1129 00:02:07.212309 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:02:07 crc kubenswrapper[4931]: E1129 00:02:07.214783 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 29 00:02:07 crc kubenswrapper[4931]: E1129 00:02:07.215003 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 29 00:02:07 crc kubenswrapper[4931]: E1129 00:02:07.215243 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7dwtb" podUID="5e38ad69-f5fb-415d-ad76-e899aed032a6" Nov 29 00:02:08 crc kubenswrapper[4931]: I1129 00:02:08.211659 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:02:08 crc kubenswrapper[4931]: I1129 00:02:08.213859 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 29 00:02:08 crc kubenswrapper[4931]: I1129 00:02:08.214761 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 29 00:02:09 crc kubenswrapper[4931]: I1129 00:02:09.211216 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:02:09 crc kubenswrapper[4931]: I1129 00:02:09.212114 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:02:09 crc kubenswrapper[4931]: I1129 00:02:09.212162 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:02:09 crc kubenswrapper[4931]: I1129 00:02:09.217050 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 29 00:02:09 crc kubenswrapper[4931]: I1129 00:02:09.217451 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 29 00:02:09 crc kubenswrapper[4931]: I1129 00:02:09.219176 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 29 00:02:09 crc kubenswrapper[4931]: I1129 00:02:09.222796 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.387001 4931 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.440370 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-lfh9z"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.440998 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.447660 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.448244 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.448335 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.447860 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.448625 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.449149 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.453312 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.453862 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.455969 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.456386 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.457387 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.457901 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.458280 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-nlp5n"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.459265 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.459354 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-nlp5n" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.459461 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.459987 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.464635 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.464789 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.464940 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.464640 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.465621 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.465886 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.466407 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.466523 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.466552 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.466667 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.466790 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.469930 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ffwsb"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.470710 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.470913 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.471145 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.471910 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-f2666"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.472363 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.472514 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.472614 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.473091 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kshsm"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.473755 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.474502 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-j4wm9"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.474961 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-j4wm9" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.483572 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.484107 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.487427 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.488224 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.494429 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.494635 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.494834 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.495135 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.497622 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.497989 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.498731 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.498773 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.498955 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.499186 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.499286 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.499399 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.499796 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.499977 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.500096 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.500490 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.501084 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.501470 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.501653 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.501862 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.502173 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.509150 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29406240-4vs86"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.509857 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29406240-4vs86" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.510213 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.510397 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.510980 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.511158 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.520565 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.520754 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.520891 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.521047 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.521171 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.521215 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.521383 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.521443 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.521480 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.521582 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.521592 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.523203 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-config\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.523250 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.523293 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h47wp\" (UniqueName: \"kubernetes.io/projected/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-kube-api-access-h47wp\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.523316 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-serving-cert\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.523351 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-service-ca-bundle\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.523371 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.524073 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.524279 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-vvmnt"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.526130 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.526951 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.526976 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.527372 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.527439 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.533307 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.549499 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.549845 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.550314 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.557668 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.558322 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xl2bw"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.558910 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.558967 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.559199 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.559404 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.559437 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.559514 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.561999 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.562455 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.562610 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.562866 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.563095 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.564494 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.565117 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.565386 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.565534 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.567529 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.568520 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.568771 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.569060 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.569352 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.570349 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.571046 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.571147 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.571210 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.571622 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rx5zl"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.572054 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.572317 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.572889 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.573053 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.573127 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.573148 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.573179 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.573333 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.573415 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.573705 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8j97r"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.574515 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.577247 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.577519 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.578929 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.579333 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.579643 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.591734 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.593799 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.595137 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-lfh9z"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.595172 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9v6hx"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.600339 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.601228 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9v6hx" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.601645 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.602699 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.603851 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.607578 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.619856 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.622755 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.623061 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.625464 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626060 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626692 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626718 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7297960-2485-4860-8587-04ebdf29ad6a-audit-dir\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626736 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvf52\" (UniqueName: \"kubernetes.io/projected/e49dd8f4-f866-4878-b867-d4d64de070c7-kube-api-access-pvf52\") pod \"downloads-7954f5f757-j4wm9\" (UID: \"e49dd8f4-f866-4878-b867-d4d64de070c7\") " pod="openshift-console/downloads-7954f5f757-j4wm9" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626759 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c7297960-2485-4860-8587-04ebdf29ad6a-audit-policies\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626776 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626795 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f6ac7072-d5d5-4c7a-8ea1-89090be1fb72-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lpxjw\" (UID: \"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626844 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnsdv\" (UniqueName: \"kubernetes.io/projected/f6ac7072-d5d5-4c7a-8ea1-89090be1fb72-kube-api-access-nnsdv\") pod \"cluster-image-registry-operator-dc59b4c8b-lpxjw\" (UID: \"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626865 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/695bbf28-25a7-4ba8-a262-fc333dd7707a-serving-cert\") pod \"openshift-config-operator-7777fb866f-vbd6v\" (UID: \"695bbf28-25a7-4ba8-a262-fc333dd7707a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626882 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0be48302-a154-4099-b4d7-3e4781964c7d-client-ca\") pod \"route-controller-manager-6576b87f9c-rqlwf\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626902 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7ac25448-7ec8-4cb7-8f19-1dab208323a4-metrics-tls\") pod \"dns-operator-744455d44c-nlp5n\" (UID: \"7ac25448-7ec8-4cb7-8f19-1dab208323a4\") " pod="openshift-dns-operator/dns-operator-744455d44c-nlp5n" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626931 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-serving-cert\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626945 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f6ac7072-d5d5-4c7a-8ea1-89090be1fb72-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lpxjw\" (UID: \"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626960 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7297960-2485-4860-8587-04ebdf29ad6a-serving-cert\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626975 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/199fba37-4269-4fb2-9249-b384fc2f2905-audit-dir\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.626991 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627008 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627023 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/695bbf28-25a7-4ba8-a262-fc333dd7707a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vbd6v\" (UID: \"695bbf28-25a7-4ba8-a262-fc333dd7707a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627039 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc24g\" (UniqueName: \"kubernetes.io/projected/695bbf28-25a7-4ba8-a262-fc333dd7707a-kube-api-access-hc24g\") pod \"openshift-config-operator-7777fb866f-vbd6v\" (UID: \"695bbf28-25a7-4ba8-a262-fc333dd7707a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627054 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627088 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft4tc\" (UniqueName: \"kubernetes.io/projected/35311d49-bd1d-4351-90a8-f332a6c3a5d2-kube-api-access-ft4tc\") pod \"image-pruner-29406240-4vs86\" (UID: \"35311d49-bd1d-4351-90a8-f332a6c3a5d2\") " pod="openshift-image-registry/image-pruner-29406240-4vs86" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627112 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627130 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-config\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627144 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7297960-2485-4860-8587-04ebdf29ad6a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627161 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-config\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627176 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79k7b\" (UniqueName: \"kubernetes.io/projected/7ac25448-7ec8-4cb7-8f19-1dab208323a4-kube-api-access-79k7b\") pod \"dns-operator-744455d44c-nlp5n\" (UID: \"7ac25448-7ec8-4cb7-8f19-1dab208323a4\") " pod="openshift-dns-operator/dns-operator-744455d44c-nlp5n" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627195 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c5e5b6b-ab0b-4158-9f5a-b6630b901e60-serving-cert\") pod \"console-operator-58897d9998-f2666\" (UID: \"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60\") " pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627210 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-config\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627228 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7297960-2485-4860-8587-04ebdf29ad6a-etcd-client\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627243 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/321d4bf6-7bff-4513-87e5-adce01251a0c-machine-approver-tls\") pod \"machine-approver-56656f9798-9c86t\" (UID: \"321d4bf6-7bff-4513-87e5-adce01251a0c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627258 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86nlr\" (UniqueName: \"kubernetes.io/projected/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-kube-api-access-86nlr\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627272 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c5e5b6b-ab0b-4158-9f5a-b6630b901e60-config\") pod \"console-operator-58897d9998-f2666\" (UID: \"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60\") " pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627287 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-audit-policies\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627302 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627323 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627341 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h47wp\" (UniqueName: \"kubernetes.io/projected/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-kube-api-access-h47wp\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627358 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-client-ca\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627373 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627387 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbgrn\" (UniqueName: \"kubernetes.io/projected/13b49f2e-d3eb-483c-a910-202c955bfda9-kube-api-access-tbgrn\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627417 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-service-ca-bundle\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627433 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627447 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/321d4bf6-7bff-4513-87e5-adce01251a0c-config\") pod \"machine-approver-56656f9798-9c86t\" (UID: \"321d4bf6-7bff-4513-87e5-adce01251a0c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627462 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-trusted-ca-bundle\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627478 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627493 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b40393fc-4539-4caa-a76c-6eda95f7483c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzbr7\" (UID: \"b40393fc-4539-4caa-a76c-6eda95f7483c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627507 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13b49f2e-d3eb-483c-a910-202c955bfda9-serving-cert\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627522 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgjgr\" (UniqueName: \"kubernetes.io/projected/b40393fc-4539-4caa-a76c-6eda95f7483c-kube-api-access-pgjgr\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzbr7\" (UID: \"b40393fc-4539-4caa-a76c-6eda95f7483c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627538 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tw99\" (UniqueName: \"kubernetes.io/projected/199fba37-4269-4fb2-9249-b384fc2f2905-kube-api-access-6tw99\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627554 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2vmm\" (UniqueName: \"kubernetes.io/projected/0be48302-a154-4099-b4d7-3e4781964c7d-kube-api-access-v2vmm\") pod \"route-controller-manager-6576b87f9c-rqlwf\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627576 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lngw\" (UniqueName: \"kubernetes.io/projected/3c5e5b6b-ab0b-4158-9f5a-b6630b901e60-kube-api-access-7lngw\") pod \"console-operator-58897d9998-f2666\" (UID: \"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60\") " pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627592 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627606 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be48302-a154-4099-b4d7-3e4781964c7d-config\") pod \"route-controller-manager-6576b87f9c-rqlwf\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627620 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7297960-2485-4860-8587-04ebdf29ad6a-encryption-config\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627635 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f6ac7072-d5d5-4c7a-8ea1-89090be1fb72-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lpxjw\" (UID: \"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627649 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpp7p\" (UniqueName: \"kubernetes.io/projected/831e6629-6933-4be3-8d45-a4a9504387d1-kube-api-access-jpp7p\") pod \"cluster-samples-operator-665b6dd947-h25zx\" (UID: \"831e6629-6933-4be3-8d45-a4a9504387d1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627666 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/321d4bf6-7bff-4513-87e5-adce01251a0c-auth-proxy-config\") pod \"machine-approver-56656f9798-9c86t\" (UID: \"321d4bf6-7bff-4513-87e5-adce01251a0c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627682 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/35311d49-bd1d-4351-90a8-f332a6c3a5d2-serviceca\") pod \"image-pruner-29406240-4vs86\" (UID: \"35311d49-bd1d-4351-90a8-f332a6c3a5d2\") " pod="openshift-image-registry/image-pruner-29406240-4vs86" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627696 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/831e6629-6933-4be3-8d45-a4a9504387d1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-h25zx\" (UID: \"831e6629-6933-4be3-8d45-a4a9504387d1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627711 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b6h5\" (UniqueName: \"kubernetes.io/projected/321d4bf6-7bff-4513-87e5-adce01251a0c-kube-api-access-4b6h5\") pod \"machine-approver-56656f9798-9c86t\" (UID: \"321d4bf6-7bff-4513-87e5-adce01251a0c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627725 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjmf8\" (UniqueName: \"kubernetes.io/projected/c7297960-2485-4860-8587-04ebdf29ad6a-kube-api-access-vjmf8\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627771 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-serving-cert\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627788 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-service-ca\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627802 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-oauth-serving-cert\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627827 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627842 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7297960-2485-4860-8587-04ebdf29ad6a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627858 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-oauth-config\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627873 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3c5e5b6b-ab0b-4158-9f5a-b6630b901e60-trusted-ca\") pod \"console-operator-58897d9998-f2666\" (UID: \"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60\") " pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627890 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0be48302-a154-4099-b4d7-3e4781964c7d-serving-cert\") pod \"route-controller-manager-6576b87f9c-rqlwf\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.627903 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b40393fc-4539-4caa-a76c-6eda95f7483c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzbr7\" (UID: \"b40393fc-4539-4caa-a76c-6eda95f7483c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.628087 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pr87s"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.628313 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.628568 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t9jm9"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.629130 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.629468 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.629495 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-service-ca-bundle\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.630095 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.630341 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-t9jm9" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.630526 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-config\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.630616 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.632603 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-plmcw"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.633403 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.636133 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ffwsb"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.638117 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.638703 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.640050 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.641193 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.641691 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.642394 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.642556 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.643137 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.643360 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-nlp5n"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.644326 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.644688 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.647151 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.647770 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kshsm"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.647919 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.655790 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-serving-cert\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.665489 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.666058 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-j4wm9"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.666113 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t9jm9"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.666130 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7c58s"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.666830 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.669407 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.670363 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.671135 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.671677 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.672044 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.672249 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.672629 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.674964 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.675622 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4zs2z"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.675982 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.676063 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.676860 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.679903 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-hbmrd"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.681294 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rx5zl"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.681419 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.682414 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vvmnt"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.683433 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.684575 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.687269 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.689919 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8j97r"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.691405 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.693955 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.694558 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.701419 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.703305 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.703616 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.704529 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-f2666"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.705480 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29406240-4vs86"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.708033 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.710098 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.712353 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.712604 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.715275 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4zs2z"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.716658 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9v6hx"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.718053 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-2sv7b"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.718649 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2sv7b" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.719465 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-8sbd2"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.720048 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8sbd2" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.720893 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.722604 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xl2bw"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.722710 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.723672 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.725103 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.726527 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pr87s"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.727981 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7c58s"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728248 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728297 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7297960-2485-4860-8587-04ebdf29ad6a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728320 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjmf8\" (UniqueName: \"kubernetes.io/projected/c7297960-2485-4860-8587-04ebdf29ad6a-kube-api-access-vjmf8\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728341 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-serving-cert\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728362 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-service-ca\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728381 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-oauth-serving-cert\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728401 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-oauth-config\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728422 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3c5e5b6b-ab0b-4158-9f5a-b6630b901e60-trusted-ca\") pod \"console-operator-58897d9998-f2666\" (UID: \"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60\") " pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728442 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b40393fc-4539-4caa-a76c-6eda95f7483c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzbr7\" (UID: \"b40393fc-4539-4caa-a76c-6eda95f7483c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728468 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/127076b5-1262-4859-837f-c1f844eafebf-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bs2jq\" (UID: \"127076b5-1262-4859-837f-c1f844eafebf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728490 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0be48302-a154-4099-b4d7-3e4781964c7d-serving-cert\") pod \"route-controller-manager-6576b87f9c-rqlwf\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728510 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728531 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7297960-2485-4860-8587-04ebdf29ad6a-audit-dir\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728552 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c7ebccc-ceef-4268-8143-5318f56e5337-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-njwv4\" (UID: \"2c7ebccc-ceef-4268-8143-5318f56e5337\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728573 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnsdv\" (UniqueName: \"kubernetes.io/projected/f6ac7072-d5d5-4c7a-8ea1-89090be1fb72-kube-api-access-nnsdv\") pod \"cluster-image-registry-operator-dc59b4c8b-lpxjw\" (UID: \"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728592 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/695bbf28-25a7-4ba8-a262-fc333dd7707a-serving-cert\") pod \"openshift-config-operator-7777fb866f-vbd6v\" (UID: \"695bbf28-25a7-4ba8-a262-fc333dd7707a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728612 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvf52\" (UniqueName: \"kubernetes.io/projected/e49dd8f4-f866-4878-b867-d4d64de070c7-kube-api-access-pvf52\") pod \"downloads-7954f5f757-j4wm9\" (UID: \"e49dd8f4-f866-4878-b867-d4d64de070c7\") " pod="openshift-console/downloads-7954f5f757-j4wm9" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728631 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c7297960-2485-4860-8587-04ebdf29ad6a-audit-policies\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728653 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728672 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f6ac7072-d5d5-4c7a-8ea1-89090be1fb72-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lpxjw\" (UID: \"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728694 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd34298a-7f8c-4264-97d5-4503c5182e49-service-ca-bundle\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728715 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0be48302-a154-4099-b4d7-3e4781964c7d-client-ca\") pod \"route-controller-manager-6576b87f9c-rqlwf\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728735 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7ac25448-7ec8-4cb7-8f19-1dab208323a4-metrics-tls\") pod \"dns-operator-744455d44c-nlp5n\" (UID: \"7ac25448-7ec8-4cb7-8f19-1dab208323a4\") " pod="openshift-dns-operator/dns-operator-744455d44c-nlp5n" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728797 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f6ac7072-d5d5-4c7a-8ea1-89090be1fb72-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lpxjw\" (UID: \"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728833 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7297960-2485-4860-8587-04ebdf29ad6a-serving-cert\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728856 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/199fba37-4269-4fb2-9249-b384fc2f2905-audit-dir\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728877 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728900 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/bd34298a-7f8c-4264-97d5-4503c5182e49-default-certificate\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728924 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728946 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/695bbf28-25a7-4ba8-a262-fc333dd7707a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vbd6v\" (UID: \"695bbf28-25a7-4ba8-a262-fc333dd7707a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728968 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc24g\" (UniqueName: \"kubernetes.io/projected/695bbf28-25a7-4ba8-a262-fc333dd7707a-kube-api-access-hc24g\") pod \"openshift-config-operator-7777fb866f-vbd6v\" (UID: \"695bbf28-25a7-4ba8-a262-fc333dd7707a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728989 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729016 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvtkq\" (UniqueName: \"kubernetes.io/projected/bd34298a-7f8c-4264-97d5-4503c5182e49-kube-api-access-jvtkq\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729045 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft4tc\" (UniqueName: \"kubernetes.io/projected/35311d49-bd1d-4351-90a8-f332a6c3a5d2-kube-api-access-ft4tc\") pod \"image-pruner-29406240-4vs86\" (UID: \"35311d49-bd1d-4351-90a8-f332a6c3a5d2\") " pod="openshift-image-registry/image-pruner-29406240-4vs86" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729067 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cbr6\" (UniqueName: \"kubernetes.io/projected/0e2c10ac-c727-42af-80a9-6e14937b986c-kube-api-access-6cbr6\") pod \"migrator-59844c95c7-9v6hx\" (UID: \"0e2c10ac-c727-42af-80a9-6e14937b986c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9v6hx" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729097 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb8c7\" (UniqueName: \"kubernetes.io/projected/127076b5-1262-4859-837f-c1f844eafebf-kube-api-access-tb8c7\") pod \"ingress-operator-5b745b69d9-bs2jq\" (UID: \"127076b5-1262-4859-837f-c1f844eafebf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729119 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729139 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7297960-2485-4860-8587-04ebdf29ad6a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729161 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-config\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729182 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79k7b\" (UniqueName: \"kubernetes.io/projected/7ac25448-7ec8-4cb7-8f19-1dab208323a4-kube-api-access-79k7b\") pod \"dns-operator-744455d44c-nlp5n\" (UID: \"7ac25448-7ec8-4cb7-8f19-1dab208323a4\") " pod="openshift-dns-operator/dns-operator-744455d44c-nlp5n" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729205 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8zq8\" (UniqueName: \"kubernetes.io/projected/2c7ebccc-ceef-4268-8143-5318f56e5337-kube-api-access-w8zq8\") pod \"kube-storage-version-migrator-operator-b67b599dd-njwv4\" (UID: \"2c7ebccc-ceef-4268-8143-5318f56e5337\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729228 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/bd34298a-7f8c-4264-97d5-4503c5182e49-stats-auth\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729250 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c5e5b6b-ab0b-4158-9f5a-b6630b901e60-serving-cert\") pod \"console-operator-58897d9998-f2666\" (UID: \"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60\") " pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729278 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-config\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729306 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c5e5b6b-ab0b-4158-9f5a-b6630b901e60-config\") pod \"console-operator-58897d9998-f2666\" (UID: \"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60\") " pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729327 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-audit-policies\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729347 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7297960-2485-4860-8587-04ebdf29ad6a-etcd-client\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729367 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/321d4bf6-7bff-4513-87e5-adce01251a0c-machine-approver-tls\") pod \"machine-approver-56656f9798-9c86t\" (UID: \"321d4bf6-7bff-4513-87e5-adce01251a0c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729388 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86nlr\" (UniqueName: \"kubernetes.io/projected/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-kube-api-access-86nlr\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729413 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/127076b5-1262-4859-837f-c1f844eafebf-trusted-ca\") pod \"ingress-operator-5b745b69d9-bs2jq\" (UID: \"127076b5-1262-4859-837f-c1f844eafebf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729437 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729465 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c7ebccc-ceef-4268-8143-5318f56e5337-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-njwv4\" (UID: \"2c7ebccc-ceef-4268-8143-5318f56e5337\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729485 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd34298a-7f8c-4264-97d5-4503c5182e49-metrics-certs\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729513 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729535 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbgrn\" (UniqueName: \"kubernetes.io/projected/13b49f2e-d3eb-483c-a910-202c955bfda9-kube-api-access-tbgrn\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729555 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-client-ca\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729668 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729691 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/321d4bf6-7bff-4513-87e5-adce01251a0c-config\") pod \"machine-approver-56656f9798-9c86t\" (UID: \"321d4bf6-7bff-4513-87e5-adce01251a0c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729713 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/127076b5-1262-4859-837f-c1f844eafebf-metrics-tls\") pod \"ingress-operator-5b745b69d9-bs2jq\" (UID: \"127076b5-1262-4859-837f-c1f844eafebf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729743 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-trusted-ca-bundle\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729763 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729783 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b40393fc-4539-4caa-a76c-6eda95f7483c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzbr7\" (UID: \"b40393fc-4539-4caa-a76c-6eda95f7483c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729802 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tw99\" (UniqueName: \"kubernetes.io/projected/199fba37-4269-4fb2-9249-b384fc2f2905-kube-api-access-6tw99\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729845 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2vmm\" (UniqueName: \"kubernetes.io/projected/0be48302-a154-4099-b4d7-3e4781964c7d-kube-api-access-v2vmm\") pod \"route-controller-manager-6576b87f9c-rqlwf\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729865 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13b49f2e-d3eb-483c-a910-202c955bfda9-serving-cert\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729873 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729888 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgjgr\" (UniqueName: \"kubernetes.io/projected/b40393fc-4539-4caa-a76c-6eda95f7483c-kube-api-access-pgjgr\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzbr7\" (UID: \"b40393fc-4539-4caa-a76c-6eda95f7483c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729900 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3c5e5b6b-ab0b-4158-9f5a-b6630b901e60-trusted-ca\") pod \"console-operator-58897d9998-f2666\" (UID: \"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60\") " pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729929 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be48302-a154-4099-b4d7-3e4781964c7d-config\") pod \"route-controller-manager-6576b87f9c-rqlwf\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729953 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7297960-2485-4860-8587-04ebdf29ad6a-encryption-config\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.729985 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lngw\" (UniqueName: \"kubernetes.io/projected/3c5e5b6b-ab0b-4158-9f5a-b6630b901e60-kube-api-access-7lngw\") pod \"console-operator-58897d9998-f2666\" (UID: \"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60\") " pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.730007 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.730036 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bfa259bd-b798-4cd3-9560-10528387a9b2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xdssh\" (UID: \"bfa259bd-b798-4cd3-9560-10528387a9b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.730057 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f6ac7072-d5d5-4c7a-8ea1-89090be1fb72-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lpxjw\" (UID: \"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.730077 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpp7p\" (UniqueName: \"kubernetes.io/projected/831e6629-6933-4be3-8d45-a4a9504387d1-kube-api-access-jpp7p\") pod \"cluster-samples-operator-665b6dd947-h25zx\" (UID: \"831e6629-6933-4be3-8d45-a4a9504387d1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.730104 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/321d4bf6-7bff-4513-87e5-adce01251a0c-auth-proxy-config\") pod \"machine-approver-56656f9798-9c86t\" (UID: \"321d4bf6-7bff-4513-87e5-adce01251a0c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.730120 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw4zr\" (UniqueName: \"kubernetes.io/projected/bfa259bd-b798-4cd3-9560-10528387a9b2-kube-api-access-rw4zr\") pod \"package-server-manager-789f6589d5-xdssh\" (UID: \"bfa259bd-b798-4cd3-9560-10528387a9b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.730138 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/35311d49-bd1d-4351-90a8-f332a6c3a5d2-serviceca\") pod \"image-pruner-29406240-4vs86\" (UID: \"35311d49-bd1d-4351-90a8-f332a6c3a5d2\") " pod="openshift-image-registry/image-pruner-29406240-4vs86" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.730154 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/831e6629-6933-4be3-8d45-a4a9504387d1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-h25zx\" (UID: \"831e6629-6933-4be3-8d45-a4a9504387d1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.730170 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b6h5\" (UniqueName: \"kubernetes.io/projected/321d4bf6-7bff-4513-87e5-adce01251a0c-kube-api-access-4b6h5\") pod \"machine-approver-56656f9798-9c86t\" (UID: \"321d4bf6-7bff-4513-87e5-adce01251a0c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.730321 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7297960-2485-4860-8587-04ebdf29ad6a-audit-dir\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.730881 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0be48302-a154-4099-b4d7-3e4781964c7d-client-ca\") pod \"route-controller-manager-6576b87f9c-rqlwf\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.731302 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c7297960-2485-4860-8587-04ebdf29ad6a-audit-policies\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.731866 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-oauth-serving-cert\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.731905 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0be48302-a154-4099-b4d7-3e4781964c7d-serving-cert\") pod \"route-controller-manager-6576b87f9c-rqlwf\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.732121 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.732282 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.732322 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2sv7b"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.732339 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.732434 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.732918 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-service-ca\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.732935 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.732977 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/199fba37-4269-4fb2-9249-b384fc2f2905-audit-dir\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.733186 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be48302-a154-4099-b4d7-3e4781964c7d-config\") pod \"route-controller-manager-6576b87f9c-rqlwf\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.733421 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7297960-2485-4860-8587-04ebdf29ad6a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.733587 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/695bbf28-25a7-4ba8-a262-fc333dd7707a-serving-cert\") pod \"openshift-config-operator-7777fb866f-vbd6v\" (UID: \"695bbf28-25a7-4ba8-a262-fc333dd7707a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.733877 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f6ac7072-d5d5-4c7a-8ea1-89090be1fb72-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lpxjw\" (UID: \"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.733976 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-serving-cert\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.734077 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-config\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.734977 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7297960-2485-4860-8587-04ebdf29ad6a-serving-cert\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.735013 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7ac25448-7ec8-4cb7-8f19-1dab208323a4-metrics-tls\") pod \"dns-operator-744455d44c-nlp5n\" (UID: \"7ac25448-7ec8-4cb7-8f19-1dab208323a4\") " pod="openshift-dns-operator/dns-operator-744455d44c-nlp5n" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.735218 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/695bbf28-25a7-4ba8-a262-fc333dd7707a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vbd6v\" (UID: \"695bbf28-25a7-4ba8-a262-fc333dd7707a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.728902 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7297960-2485-4860-8587-04ebdf29ad6a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.736183 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/321d4bf6-7bff-4513-87e5-adce01251a0c-auth-proxy-config\") pod \"machine-approver-56656f9798-9c86t\" (UID: \"321d4bf6-7bff-4513-87e5-adce01251a0c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.736872 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/35311d49-bd1d-4351-90a8-f332a6c3a5d2-serviceca\") pod \"image-pruner-29406240-4vs86\" (UID: \"35311d49-bd1d-4351-90a8-f332a6c3a5d2\") " pod="openshift-image-registry/image-pruner-29406240-4vs86" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.737233 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.737270 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.737396 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c5e5b6b-ab0b-4158-9f5a-b6630b901e60-config\") pod \"console-operator-58897d9998-f2666\" (UID: \"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60\") " pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.737432 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-audit-policies\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.737976 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c5e5b6b-ab0b-4158-9f5a-b6630b901e60-serving-cert\") pod \"console-operator-58897d9998-f2666\" (UID: \"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60\") " pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.737985 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b40393fc-4539-4caa-a76c-6eda95f7483c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzbr7\" (UID: \"b40393fc-4539-4caa-a76c-6eda95f7483c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.737990 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.738246 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.738248 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-trusted-ca-bundle\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.738336 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/321d4bf6-7bff-4513-87e5-adce01251a0c-config\") pod \"machine-approver-56656f9798-9c86t\" (UID: \"321d4bf6-7bff-4513-87e5-adce01251a0c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.738505 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-client-ca\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.738685 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b40393fc-4539-4caa-a76c-6eda95f7483c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzbr7\" (UID: \"b40393fc-4539-4caa-a76c-6eda95f7483c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.739080 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-oauth-config\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.739084 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7297960-2485-4860-8587-04ebdf29ad6a-encryption-config\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.739437 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.739787 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13b49f2e-d3eb-483c-a910-202c955bfda9-serving-cert\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.739841 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hbmrd"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.740582 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f6ac7072-d5d5-4c7a-8ea1-89090be1fb72-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lpxjw\" (UID: \"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.740596 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.740641 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7297960-2485-4860-8587-04ebdf29ad6a-etcd-client\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.740930 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.740966 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-p4jwd"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.741326 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/321d4bf6-7bff-4513-87e5-adce01251a0c-machine-approver-tls\") pod \"machine-approver-56656f9798-9c86t\" (UID: \"321d4bf6-7bff-4513-87e5-adce01251a0c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.741868 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-p4jwd"] Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.741926 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.741941 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.742130 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.742856 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.743478 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/831e6629-6933-4be3-8d45-a4a9504387d1-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-h25zx\" (UID: \"831e6629-6933-4be3-8d45-a4a9504387d1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.745390 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.746089 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-config\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.763171 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.783767 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.802731 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.823304 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.830752 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/bd34298a-7f8c-4264-97d5-4503c5182e49-default-certificate\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.830905 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvtkq\" (UniqueName: \"kubernetes.io/projected/bd34298a-7f8c-4264-97d5-4503c5182e49-kube-api-access-jvtkq\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.831091 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cbr6\" (UniqueName: \"kubernetes.io/projected/0e2c10ac-c727-42af-80a9-6e14937b986c-kube-api-access-6cbr6\") pod \"migrator-59844c95c7-9v6hx\" (UID: \"0e2c10ac-c727-42af-80a9-6e14937b986c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9v6hx" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.831596 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb8c7\" (UniqueName: \"kubernetes.io/projected/127076b5-1262-4859-837f-c1f844eafebf-kube-api-access-tb8c7\") pod \"ingress-operator-5b745b69d9-bs2jq\" (UID: \"127076b5-1262-4859-837f-c1f844eafebf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.831730 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8zq8\" (UniqueName: \"kubernetes.io/projected/2c7ebccc-ceef-4268-8143-5318f56e5337-kube-api-access-w8zq8\") pod \"kube-storage-version-migrator-operator-b67b599dd-njwv4\" (UID: \"2c7ebccc-ceef-4268-8143-5318f56e5337\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.831824 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/bd34298a-7f8c-4264-97d5-4503c5182e49-stats-auth\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.831905 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/127076b5-1262-4859-837f-c1f844eafebf-trusted-ca\") pod \"ingress-operator-5b745b69d9-bs2jq\" (UID: \"127076b5-1262-4859-837f-c1f844eafebf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.831986 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c7ebccc-ceef-4268-8143-5318f56e5337-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-njwv4\" (UID: \"2c7ebccc-ceef-4268-8143-5318f56e5337\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.832065 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd34298a-7f8c-4264-97d5-4503c5182e49-metrics-certs\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.832608 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/127076b5-1262-4859-837f-c1f844eafebf-metrics-tls\") pod \"ingress-operator-5b745b69d9-bs2jq\" (UID: \"127076b5-1262-4859-837f-c1f844eafebf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.832832 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bfa259bd-b798-4cd3-9560-10528387a9b2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xdssh\" (UID: \"bfa259bd-b798-4cd3-9560-10528387a9b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.832958 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw4zr\" (UniqueName: \"kubernetes.io/projected/bfa259bd-b798-4cd3-9560-10528387a9b2-kube-api-access-rw4zr\") pod \"package-server-manager-789f6589d5-xdssh\" (UID: \"bfa259bd-b798-4cd3-9560-10528387a9b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.833091 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/127076b5-1262-4859-837f-c1f844eafebf-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bs2jq\" (UID: \"127076b5-1262-4859-837f-c1f844eafebf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.833205 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c7ebccc-ceef-4268-8143-5318f56e5337-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-njwv4\" (UID: \"2c7ebccc-ceef-4268-8143-5318f56e5337\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.833336 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd34298a-7f8c-4264-97d5-4503c5182e49-service-ca-bundle\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.833740 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/127076b5-1262-4859-837f-c1f844eafebf-trusted-ca\") pod \"ingress-operator-5b745b69d9-bs2jq\" (UID: \"127076b5-1262-4859-837f-c1f844eafebf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.835720 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/127076b5-1262-4859-837f-c1f844eafebf-metrics-tls\") pod \"ingress-operator-5b745b69d9-bs2jq\" (UID: \"127076b5-1262-4859-837f-c1f844eafebf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.843330 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.870627 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.884210 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.903463 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.922843 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.942691 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.964574 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 29 00:02:14 crc kubenswrapper[4931]: I1129 00:02:14.983386 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.004453 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.038550 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.039166 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.039334 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.039779 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.040185 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.041043 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:02:15 crc kubenswrapper[4931]: E1129 00:02:15.041237 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:04:17.041213201 +0000 UTC m=+270.203106443 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.045110 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.049394 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.050064 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.050616 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.063670 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.083151 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.103946 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.124208 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.136461 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.143618 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.163569 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.184352 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.204595 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.224005 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.239860 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.244625 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.250785 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.264303 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.304367 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.323642 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.324449 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h47wp\" (UniqueName: \"kubernetes.io/projected/e27eaa50-e43a-4882-8bb3-ec3624d6da2d-kube-api-access-h47wp\") pod \"authentication-operator-69f744f599-lfh9z\" (UID: \"e27eaa50-e43a-4882-8bb3-ec3624d6da2d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.343498 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.366776 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.372544 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.385468 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.402712 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.423477 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.442636 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.462879 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.479035 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/bd34298a-7f8c-4264-97d5-4503c5182e49-default-certificate\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.483506 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.503882 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.523551 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.538208 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd34298a-7f8c-4264-97d5-4503c5182e49-metrics-certs\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.544681 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.570224 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.577687 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/bd34298a-7f8c-4264-97d5-4503c5182e49-stats-auth\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.584240 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.597611 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd34298a-7f8c-4264-97d5-4503c5182e49-service-ca-bundle\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.602135 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-lfh9z"] Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.604172 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 29 00:02:15 crc kubenswrapper[4931]: W1129 00:02:15.614153 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode27eaa50_e43a_4882_8bb3_ec3624d6da2d.slice/crio-0a9d23e1ecd3a0ec34de9cd28451f1bbd84a2e3ee3e9d2765196b3e66acd2b51 WatchSource:0}: Error finding container 0a9d23e1ecd3a0ec34de9cd28451f1bbd84a2e3ee3e9d2765196b3e66acd2b51: Status 404 returned error can't find the container with id 0a9d23e1ecd3a0ec34de9cd28451f1bbd84a2e3ee3e9d2765196b3e66acd2b51 Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.622927 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.642275 4931 request.go:700] Waited for 1.003277683s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dserving-cert&limit=500&resourceVersion=0 Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.643297 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.646317 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c7ebccc-ceef-4268-8143-5318f56e5337-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-njwv4\" (UID: \"2c7ebccc-ceef-4268-8143-5318f56e5337\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.662682 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.682518 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.703324 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.723344 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.743135 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.764235 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.783331 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.803404 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.823793 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 29 00:02:15 crc kubenswrapper[4931]: E1129 00:02:15.833431 4931 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 29 00:02:15 crc kubenswrapper[4931]: E1129 00:02:15.833535 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bfa259bd-b798-4cd3-9560-10528387a9b2-package-server-manager-serving-cert podName:bfa259bd-b798-4cd3-9560-10528387a9b2 nodeName:}" failed. No retries permitted until 2025-11-29 00:02:16.333507543 +0000 UTC m=+149.495400805 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/bfa259bd-b798-4cd3-9560-10528387a9b2-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-xdssh" (UID: "bfa259bd-b798-4cd3-9560-10528387a9b2") : failed to sync secret cache: timed out waiting for the condition Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.843321 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.864049 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.883697 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.905043 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.944348 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.964239 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 29 00:02:15 crc kubenswrapper[4931]: I1129 00:02:15.984040 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.003218 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.025179 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.043714 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.064666 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.076638 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"636883d00acd0832f997bb57eb077cff7c120a43113b11a0f57d547d4f2d8fd1"} Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.078359 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" event={"ID":"e27eaa50-e43a-4882-8bb3-ec3624d6da2d","Type":"ContainerStarted","Data":"0a9d23e1ecd3a0ec34de9cd28451f1bbd84a2e3ee3e9d2765196b3e66acd2b51"} Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.080135 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7dc1266da14fa9f3b064e0d495b4e49ba59c2f5ea4dfcf3c61fddd342ffa11f4"} Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.081473 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c77296f4b1f652b89d0ef9907bb0f24a768b300c389ab05072fd871062930224"} Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.083730 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.103495 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.124341 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.145062 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.174941 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.183675 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.203846 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.216146 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c7ebccc-ceef-4268-8143-5318f56e5337-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-njwv4\" (UID: \"2c7ebccc-ceef-4268-8143-5318f56e5337\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.224343 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.243855 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.264072 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.285136 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.304265 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.324167 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.345338 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.359570 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bfa259bd-b798-4cd3-9560-10528387a9b2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xdssh\" (UID: \"bfa259bd-b798-4cd3-9560-10528387a9b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.366332 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bfa259bd-b798-4cd3-9560-10528387a9b2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xdssh\" (UID: \"bfa259bd-b798-4cd3-9560-10528387a9b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.368636 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.384118 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.404082 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.424223 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.443566 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.464629 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.483867 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.504024 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.538089 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnsdv\" (UniqueName: \"kubernetes.io/projected/f6ac7072-d5d5-4c7a-8ea1-89090be1fb72-kube-api-access-nnsdv\") pod \"cluster-image-registry-operator-dc59b4c8b-lpxjw\" (UID: \"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.566880 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgjgr\" (UniqueName: \"kubernetes.io/projected/b40393fc-4539-4caa-a76c-6eda95f7483c-kube-api-access-pgjgr\") pod \"openshift-controller-manager-operator-756b6f6bc6-hzbr7\" (UID: \"b40393fc-4539-4caa-a76c-6eda95f7483c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.577803 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjmf8\" (UniqueName: \"kubernetes.io/projected/c7297960-2485-4860-8587-04ebdf29ad6a-kube-api-access-vjmf8\") pod \"apiserver-7bbb656c7d-ncrdv\" (UID: \"c7297960-2485-4860-8587-04ebdf29ad6a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.585340 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.605542 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b6h5\" (UniqueName: \"kubernetes.io/projected/321d4bf6-7bff-4513-87e5-adce01251a0c-kube-api-access-4b6h5\") pod \"machine-approver-56656f9798-9c86t\" (UID: \"321d4bf6-7bff-4513-87e5-adce01251a0c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.619360 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvf52\" (UniqueName: \"kubernetes.io/projected/e49dd8f4-f866-4878-b867-d4d64de070c7-kube-api-access-pvf52\") pod \"downloads-7954f5f757-j4wm9\" (UID: \"e49dd8f4-f866-4878-b867-d4d64de070c7\") " pod="openshift-console/downloads-7954f5f757-j4wm9" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.642854 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft4tc\" (UniqueName: \"kubernetes.io/projected/35311d49-bd1d-4351-90a8-f332a6c3a5d2-kube-api-access-ft4tc\") pod \"image-pruner-29406240-4vs86\" (UID: \"35311d49-bd1d-4351-90a8-f332a6c3a5d2\") " pod="openshift-image-registry/image-pruner-29406240-4vs86" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.666707 4931 request.go:700] Waited for 1.931603124s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-samples-operator/serviceaccounts/cluster-samples-operator/token Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.669704 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79k7b\" (UniqueName: \"kubernetes.io/projected/7ac25448-7ec8-4cb7-8f19-1dab208323a4-kube-api-access-79k7b\") pod \"dns-operator-744455d44c-nlp5n\" (UID: \"7ac25448-7ec8-4cb7-8f19-1dab208323a4\") " pod="openshift-dns-operator/dns-operator-744455d44c-nlp5n" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.687453 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpp7p\" (UniqueName: \"kubernetes.io/projected/831e6629-6933-4be3-8d45-a4a9504387d1-kube-api-access-jpp7p\") pod \"cluster-samples-operator-665b6dd947-h25zx\" (UID: \"831e6629-6933-4be3-8d45-a4a9504387d1\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.707735 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-j4wm9" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.711707 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lngw\" (UniqueName: \"kubernetes.io/projected/3c5e5b6b-ab0b-4158-9f5a-b6630b901e60-kube-api-access-7lngw\") pod \"console-operator-58897d9998-f2666\" (UID: \"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60\") " pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.713363 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.722253 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc24g\" (UniqueName: \"kubernetes.io/projected/695bbf28-25a7-4ba8-a262-fc333dd7707a-kube-api-access-hc24g\") pod \"openshift-config-operator-7777fb866f-vbd6v\" (UID: \"695bbf28-25a7-4ba8-a262-fc333dd7707a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.736325 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.752410 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f6ac7072-d5d5-4c7a-8ea1-89090be1fb72-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lpxjw\" (UID: \"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.759624 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86nlr\" (UniqueName: \"kubernetes.io/projected/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-kube-api-access-86nlr\") pod \"console-f9d7485db-vvmnt\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.767254 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29406240-4vs86" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.782038 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.783031 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tw99\" (UniqueName: \"kubernetes.io/projected/199fba37-4269-4fb2-9249-b384fc2f2905-kube-api-access-6tw99\") pod \"oauth-openshift-558db77b4-kshsm\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.789926 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.795128 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.806955 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbgrn\" (UniqueName: \"kubernetes.io/projected/13b49f2e-d3eb-483c-a910-202c955bfda9-kube-api-access-tbgrn\") pod \"controller-manager-879f6c89f-ffwsb\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.822660 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2vmm\" (UniqueName: \"kubernetes.io/projected/0be48302-a154-4099-b4d7-3e4781964c7d-kube-api-access-v2vmm\") pod \"route-controller-manager-6576b87f9c-rqlwf\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.823270 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.845456 4931 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.862298 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv"] Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.863577 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.897333 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.905947 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvtkq\" (UniqueName: \"kubernetes.io/projected/bd34298a-7f8c-4264-97d5-4503c5182e49-kube-api-access-jvtkq\") pod \"router-default-5444994796-plmcw\" (UID: \"bd34298a-7f8c-4264-97d5-4503c5182e49\") " pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.913048 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-nlp5n" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.920043 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.927667 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cbr6\" (UniqueName: \"kubernetes.io/projected/0e2c10ac-c727-42af-80a9-6e14937b986c-kube-api-access-6cbr6\") pod \"migrator-59844c95c7-9v6hx\" (UID: \"0e2c10ac-c727-42af-80a9-6e14937b986c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9v6hx" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.939544 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb8c7\" (UniqueName: \"kubernetes.io/projected/127076b5-1262-4859-837f-c1f844eafebf-kube-api-access-tb8c7\") pod \"ingress-operator-5b745b69d9-bs2jq\" (UID: \"127076b5-1262-4859-837f-c1f844eafebf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.951093 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.960739 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.962438 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8zq8\" (UniqueName: \"kubernetes.io/projected/2c7ebccc-ceef-4268-8143-5318f56e5337-kube-api-access-w8zq8\") pod \"kube-storage-version-migrator-operator-b67b599dd-njwv4\" (UID: \"2c7ebccc-ceef-4268-8143-5318f56e5337\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.980963 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:16 crc kubenswrapper[4931]: I1129 00:02:16.988898 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw4zr\" (UniqueName: \"kubernetes.io/projected/bfa259bd-b798-4cd3-9560-10528387a9b2-kube-api-access-rw4zr\") pod \"package-server-manager-789f6589d5-xdssh\" (UID: \"bfa259bd-b798-4cd3-9560-10528387a9b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.011007 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/127076b5-1262-4859-837f-c1f844eafebf-bound-sa-token\") pod \"ingress-operator-5b745b69d9-bs2jq\" (UID: \"127076b5-1262-4859-837f-c1f844eafebf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.040236 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.081445 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d716278-6e64-4e98-b111-20228be9d375-serving-cert\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.081951 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52e44859-bfe5-4b11-8f91-5722be14dbf7-config\") pod \"kube-controller-manager-operator-78b949d7b-vw8ff\" (UID: \"52e44859-bfe5-4b11-8f91-5722be14dbf7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.082058 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.082129 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/980879e9-818d-414c-97c0-ad63e6cf4039-config\") pod \"machine-api-operator-5694c8668f-xl2bw\" (UID: \"980879e9-818d-414c-97c0-ad63e6cf4039\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.082199 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92d953e3-3660-4178-a6ee-5c28d37c08fa-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8km27\" (UID: \"92d953e3-3660-4178-a6ee-5c28d37c08fa\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.082286 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02fd0443-734e-4aad-ad80-b776e53ded19-serving-cert\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: E1129 00:02:17.082588 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:17.5825773 +0000 UTC m=+150.744470532 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.082904 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bfzjb\" (UID: \"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.082978 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgkqz\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-kube-api-access-kgkqz\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083054 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ljv5\" (UniqueName: \"kubernetes.io/projected/d078a0ae-522e-4694-98e9-8cae5563a71f-kube-api-access-7ljv5\") pod \"olm-operator-6b444d44fb-c84rp\" (UID: \"d078a0ae-522e-4694-98e9-8cae5563a71f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083135 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7d716278-6e64-4e98-b111-20228be9d375-encryption-config\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083242 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93297fe9-2486-4b24-b6c6-cb917ab240c5-trusted-ca\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083346 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjgmn\" (UniqueName: \"kubernetes.io/projected/980879e9-818d-414c-97c0-ad63e6cf4039-kube-api-access-kjgmn\") pod \"machine-api-operator-5694c8668f-xl2bw\" (UID: \"980879e9-818d-414c-97c0-ad63e6cf4039\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083371 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d078a0ae-522e-4694-98e9-8cae5563a71f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-c84rp\" (UID: \"d078a0ae-522e-4694-98e9-8cae5563a71f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083404 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/980879e9-818d-414c-97c0-ad63e6cf4039-images\") pod \"machine-api-operator-5694c8668f-xl2bw\" (UID: \"980879e9-818d-414c-97c0-ad63e6cf4039\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083433 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm4bj\" (UniqueName: \"kubernetes.io/projected/10325f8c-feb6-4dee-8d8c-9ef797f51175-kube-api-access-rm4bj\") pod \"machine-config-controller-84d6567774-rqp28\" (UID: \"10325f8c-feb6-4dee-8d8c-9ef797f51175\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083481 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52e44859-bfe5-4b11-8f91-5722be14dbf7-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-vw8ff\" (UID: \"52e44859-bfe5-4b11-8f91-5722be14dbf7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083502 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92d953e3-3660-4178-a6ee-5c28d37c08fa-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8km27\" (UID: \"92d953e3-3660-4178-a6ee-5c28d37c08fa\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083517 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/93297fe9-2486-4b24-b6c6-cb917ab240c5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083536 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/93297fe9-2486-4b24-b6c6-cb917ab240c5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083564 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/02fd0443-734e-4aad-ad80-b776e53ded19-etcd-service-ca\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083585 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/980879e9-818d-414c-97c0-ad63e6cf4039-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xl2bw\" (UID: \"980879e9-818d-414c-97c0-ad63e6cf4039\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083608 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083628 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fd0443-734e-4aad-ad80-b776e53ded19-config\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083644 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f1416be-5367-4e50-8d31-3b10657f02df-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2p679\" (UID: \"9f1416be-5367-4e50-8d31-3b10657f02df\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083666 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnf2p\" (UniqueName: \"kubernetes.io/projected/9f1416be-5367-4e50-8d31-3b10657f02df-kube-api-access-rnf2p\") pod \"openshift-apiserver-operator-796bbdcf4f-2p679\" (UID: \"9f1416be-5367-4e50-8d31-3b10657f02df\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083685 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d078a0ae-522e-4694-98e9-8cae5563a71f-srv-cert\") pod \"olm-operator-6b444d44fb-c84rp\" (UID: \"d078a0ae-522e-4694-98e9-8cae5563a71f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083707 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7d716278-6e64-4e98-b111-20228be9d375-etcd-client\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083742 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae4bca1d-744f-45c1-bb86-97713834ff4d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mzdnj\" (UID: \"ae4bca1d-744f-45c1-bb86-97713834ff4d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.083768 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blq49\" (UniqueName: \"kubernetes.io/projected/7d716278-6e64-4e98-b111-20228be9d375-kube-api-access-blq49\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084461 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae4bca1d-744f-45c1-bb86-97713834ff4d-config\") pod \"kube-apiserver-operator-766d6c64bb-mzdnj\" (UID: \"ae4bca1d-744f-45c1-bb86-97713834ff4d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084530 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10325f8c-feb6-4dee-8d8c-9ef797f51175-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rqp28\" (UID: \"10325f8c-feb6-4dee-8d8c-9ef797f51175\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084569 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ac15726e-cd7f-4836-b88c-d832e321ba93-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t9jm9\" (UID: \"ac15726e-cd7f-4836-b88c-d832e321ba93\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t9jm9" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084589 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfbvr\" (UniqueName: \"kubernetes.io/projected/ac15726e-cd7f-4836-b88c-d832e321ba93-kube-api-access-vfbvr\") pod \"multus-admission-controller-857f4d67dd-t9jm9\" (UID: \"ac15726e-cd7f-4836-b88c-d832e321ba93\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t9jm9" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084605 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/02fd0443-734e-4aad-ad80-b776e53ded19-etcd-ca\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084622 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b6c59839-1a09-4b0b-90ac-eb9d001e9581-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pgslt\" (UID: \"b6c59839-1a09-4b0b-90ac-eb9d001e9581\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084713 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/92d953e3-3660-4178-a6ee-5c28d37c08fa-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8km27\" (UID: \"92d953e3-3660-4178-a6ee-5c28d37c08fa\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084734 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-etcd-serving-ca\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084752 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10325f8c-feb6-4dee-8d8c-9ef797f51175-proxy-tls\") pod \"machine-config-controller-84d6567774-rqp28\" (UID: \"10325f8c-feb6-4dee-8d8c-9ef797f51175\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084780 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c-proxy-tls\") pod \"machine-config-operator-74547568cd-bfzjb\" (UID: \"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084800 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c-images\") pod \"machine-config-operator-74547568cd-bfzjb\" (UID: \"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084832 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52e44859-bfe5-4b11-8f91-5722be14dbf7-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-vw8ff\" (UID: \"52e44859-bfe5-4b11-8f91-5722be14dbf7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084922 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f1416be-5367-4e50-8d31-3b10657f02df-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2p679\" (UID: \"9f1416be-5367-4e50-8d31-3b10657f02df\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084948 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-config\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.084970 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdttl\" (UniqueName: \"kubernetes.io/projected/02fd0443-734e-4aad-ad80-b776e53ded19-kube-api-access-bdttl\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.085004 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7d716278-6e64-4e98-b111-20228be9d375-node-pullsecrets\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.085023 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/02fd0443-734e-4aad-ad80-b776e53ded19-etcd-client\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.085041 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/93297fe9-2486-4b24-b6c6-cb917ab240c5-registry-certificates\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.085055 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-image-import-ca\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.086173 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-audit\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.086205 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7d716278-6e64-4e98-b111-20228be9d375-audit-dir\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.086225 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cb5l\" (UniqueName: \"kubernetes.io/projected/b6c59839-1a09-4b0b-90ac-eb9d001e9581-kube-api-access-9cb5l\") pod \"control-plane-machine-set-operator-78cbb6b69f-pgslt\" (UID: \"b6c59839-1a09-4b0b-90ac-eb9d001e9581\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.086256 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-bound-sa-token\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.086279 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae4bca1d-744f-45c1-bb86-97713834ff4d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mzdnj\" (UID: \"ae4bca1d-744f-45c1-bb86-97713834ff4d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.086302 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnx8p\" (UniqueName: \"kubernetes.io/projected/233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c-kube-api-access-lnx8p\") pod \"machine-config-operator-74547568cd-bfzjb\" (UID: \"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.086332 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-registry-tls\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.115932 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" event={"ID":"c7297960-2485-4860-8587-04ebdf29ad6a","Type":"ContainerStarted","Data":"647de94d6fbfb00b40fe91d59d337c13fc52b4ad87e00af05c188eb508106cb8"} Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.121542 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.128945 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"5dfae38aff13a9e94297c1f1f4435bb9db7782b5da22b03613fe43a297897693"} Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.129393 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.135638 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" event={"ID":"321d4bf6-7bff-4513-87e5-adce01251a0c","Type":"ContainerStarted","Data":"af5942ab46e406a435745aaf3007807058f5aba115e3351c4c84233b83f114cd"} Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.155091 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" event={"ID":"e27eaa50-e43a-4882-8bb3-ec3624d6da2d","Type":"ContainerStarted","Data":"37874782a6c180ecef9de50fa7731932efe2ad6e0c98dafc5f7a71e2b5519b0d"} Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.165736 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-plmcw" event={"ID":"bd34298a-7f8c-4264-97d5-4503c5182e49","Type":"ContainerStarted","Data":"067c8b4c92a461ab01a275b2972e67986cf89ce5c0dcf43c290e41e00a570a61"} Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.169742 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7481bea520afeb7363f0705bb47d803e84cd1efc8d7ff510210299e9073593cc"} Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.171002 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"84a0056a5f082699088ed85dbe6e87e72a495aa1e2f853f13ba78a58d849f0f1"} Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.180345 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9v6hx" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.182281 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.193654 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:17 crc kubenswrapper[4931]: E1129 00:02:17.193877 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:17.693857269 +0000 UTC m=+150.855750501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.193909 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c-proxy-tls\") pod \"machine-config-operator-74547568cd-bfzjb\" (UID: \"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.193929 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c-images\") pod \"machine-config-operator-74547568cd-bfzjb\" (UID: \"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.193945 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52e44859-bfe5-4b11-8f91-5722be14dbf7-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-vw8ff\" (UID: \"52e44859-bfe5-4b11-8f91-5722be14dbf7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.193965 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh895\" (UniqueName: \"kubernetes.io/projected/c772f56c-3fa9-4331-b439-ceea5f985452-kube-api-access-zh895\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.193984 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f1416be-5367-4e50-8d31-3b10657f02df-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2p679\" (UID: \"9f1416be-5367-4e50-8d31-3b10657f02df\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194003 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4zs2z\" (UID: \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194027 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-config\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194045 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdttl\" (UniqueName: \"kubernetes.io/projected/02fd0443-734e-4aad-ad80-b776e53ded19-kube-api-access-bdttl\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194061 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7d716278-6e64-4e98-b111-20228be9d375-node-pullsecrets\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194080 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab-signing-cabundle\") pod \"service-ca-9c57cc56f-7c58s\" (UID: \"dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab\") " pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194094 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/02fd0443-734e-4aad-ad80-b776e53ded19-etcd-client\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194137 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f950cada-6ee4-447d-b82f-0cfbdc3e98f7-config\") pod \"service-ca-operator-777779d784-6gv5x\" (UID: \"f950cada-6ee4-447d-b82f-0cfbdc3e98f7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194155 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/93297fe9-2486-4b24-b6c6-cb917ab240c5-registry-certificates\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194169 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-image-import-ca\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194184 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sxqq\" (UniqueName: \"kubernetes.io/projected/c6dc9fc9-525e-4508-940f-626e636bac4b-kube-api-access-2sxqq\") pod \"machine-config-server-8sbd2\" (UID: \"c6dc9fc9-525e-4508-940f-626e636bac4b\") " pod="openshift-machine-config-operator/machine-config-server-8sbd2" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194210 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f950cada-6ee4-447d-b82f-0cfbdc3e98f7-serving-cert\") pod \"service-ca-operator-777779d784-6gv5x\" (UID: \"f950cada-6ee4-447d-b82f-0cfbdc3e98f7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194226 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-audit\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194242 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7d716278-6e64-4e98-b111-20228be9d375-audit-dir\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194257 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cb5l\" (UniqueName: \"kubernetes.io/projected/b6c59839-1a09-4b0b-90ac-eb9d001e9581-kube-api-access-9cb5l\") pod \"control-plane-machine-set-operator-78cbb6b69f-pgslt\" (UID: \"b6c59839-1a09-4b0b-90ac-eb9d001e9581\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194300 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-csi-data-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194323 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9bd673da-0ab0-468d-adb3-fbc74d2a091a-apiservice-cert\") pod \"packageserver-d55dfcdfc-9xcvz\" (UID: \"9bd673da-0ab0-468d-adb3-fbc74d2a091a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194381 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-bound-sa-token\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194397 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae4bca1d-744f-45c1-bb86-97713834ff4d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mzdnj\" (UID: \"ae4bca1d-744f-45c1-bb86-97713834ff4d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194411 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9bd673da-0ab0-468d-adb3-fbc74d2a091a-tmpfs\") pod \"packageserver-d55dfcdfc-9xcvz\" (UID: \"9bd673da-0ab0-468d-adb3-fbc74d2a091a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194439 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnx8p\" (UniqueName: \"kubernetes.io/projected/233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c-kube-api-access-lnx8p\") pod \"machine-config-operator-74547568cd-bfzjb\" (UID: \"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194462 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-registry-tls\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194492 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4zs2z\" (UID: \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194507 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fcmd\" (UniqueName: \"kubernetes.io/projected/9bd673da-0ab0-468d-adb3-fbc74d2a091a-kube-api-access-6fcmd\") pod \"packageserver-d55dfcdfc-9xcvz\" (UID: \"9bd673da-0ab0-468d-adb3-fbc74d2a091a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194521 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-plugins-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194539 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d716278-6e64-4e98-b111-20228be9d375-serving-cert\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194554 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52e44859-bfe5-4b11-8f91-5722be14dbf7-config\") pod \"kube-controller-manager-operator-78b949d7b-vw8ff\" (UID: \"52e44859-bfe5-4b11-8f91-5722be14dbf7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194573 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194590 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/980879e9-818d-414c-97c0-ad63e6cf4039-config\") pod \"machine-api-operator-5694c8668f-xl2bw\" (UID: \"980879e9-818d-414c-97c0-ad63e6cf4039\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194606 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4f7k\" (UniqueName: \"kubernetes.io/projected/dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab-kube-api-access-m4f7k\") pod \"service-ca-9c57cc56f-7c58s\" (UID: \"dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab\") " pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194622 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96bbcdc9-9b57-409e-b5b9-79a99034208a-secret-volume\") pod \"collect-profiles-29406240-fjsn7\" (UID: \"96bbcdc9-9b57-409e-b5b9-79a99034208a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194656 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e9a4efdc-7efb-4f90-97f2-ba975c3b172c-profile-collector-cert\") pod \"catalog-operator-68c6474976-v8kgk\" (UID: \"e9a4efdc-7efb-4f90-97f2-ba975c3b172c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194670 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-mountpoint-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194686 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92d953e3-3660-4178-a6ee-5c28d37c08fa-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8km27\" (UID: \"92d953e3-3660-4178-a6ee-5c28d37c08fa\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194701 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02fd0443-734e-4aad-ad80-b776e53ded19-serving-cert\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194716 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bfzjb\" (UID: \"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194731 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgkqz\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-kube-api-access-kgkqz\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194747 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ljv5\" (UniqueName: \"kubernetes.io/projected/d078a0ae-522e-4694-98e9-8cae5563a71f-kube-api-access-7ljv5\") pod \"olm-operator-6b444d44fb-c84rp\" (UID: \"d078a0ae-522e-4694-98e9-8cae5563a71f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194779 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7d716278-6e64-4e98-b111-20228be9d375-encryption-config\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194802 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93297fe9-2486-4b24-b6c6-cb917ab240c5-trusted-ca\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194858 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab-signing-key\") pod \"service-ca-9c57cc56f-7c58s\" (UID: \"dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab\") " pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194873 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ck2r\" (UniqueName: \"kubernetes.io/projected/96bbcdc9-9b57-409e-b5b9-79a99034208a-kube-api-access-6ck2r\") pod \"collect-profiles-29406240-fjsn7\" (UID: \"96bbcdc9-9b57-409e-b5b9-79a99034208a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194898 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96bbcdc9-9b57-409e-b5b9-79a99034208a-config-volume\") pod \"collect-profiles-29406240-fjsn7\" (UID: \"96bbcdc9-9b57-409e-b5b9-79a99034208a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194932 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjgmn\" (UniqueName: \"kubernetes.io/projected/980879e9-818d-414c-97c0-ad63e6cf4039-kube-api-access-kjgmn\") pod \"machine-api-operator-5694c8668f-xl2bw\" (UID: \"980879e9-818d-414c-97c0-ad63e6cf4039\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194948 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d078a0ae-522e-4694-98e9-8cae5563a71f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-c84rp\" (UID: \"d078a0ae-522e-4694-98e9-8cae5563a71f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.194982 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr2jn\" (UniqueName: \"kubernetes.io/projected/e9a4efdc-7efb-4f90-97f2-ba975c3b172c-kube-api-access-tr2jn\") pod \"catalog-operator-68c6474976-v8kgk\" (UID: \"e9a4efdc-7efb-4f90-97f2-ba975c3b172c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195005 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/980879e9-818d-414c-97c0-ad63e6cf4039-images\") pod \"machine-api-operator-5694c8668f-xl2bw\" (UID: \"980879e9-818d-414c-97c0-ad63e6cf4039\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195021 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm4bj\" (UniqueName: \"kubernetes.io/projected/10325f8c-feb6-4dee-8d8c-9ef797f51175-kube-api-access-rm4bj\") pod \"machine-config-controller-84d6567774-rqp28\" (UID: \"10325f8c-feb6-4dee-8d8c-9ef797f51175\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195038 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52e44859-bfe5-4b11-8f91-5722be14dbf7-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-vw8ff\" (UID: \"52e44859-bfe5-4b11-8f91-5722be14dbf7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195053 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9bd673da-0ab0-468d-adb3-fbc74d2a091a-webhook-cert\") pod \"packageserver-d55dfcdfc-9xcvz\" (UID: \"9bd673da-0ab0-468d-adb3-fbc74d2a091a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195069 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct58h\" (UniqueName: \"kubernetes.io/projected/47377b8d-2f9c-4a58-a99b-a94111193e23-kube-api-access-ct58h\") pod \"ingress-canary-2sv7b\" (UID: \"47377b8d-2f9c-4a58-a99b-a94111193e23\") " pod="openshift-ingress-canary/ingress-canary-2sv7b" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195093 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92d953e3-3660-4178-a6ee-5c28d37c08fa-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8km27\" (UID: \"92d953e3-3660-4178-a6ee-5c28d37c08fa\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195108 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c6dc9fc9-525e-4508-940f-626e636bac4b-node-bootstrap-token\") pod \"machine-config-server-8sbd2\" (UID: \"c6dc9fc9-525e-4508-940f-626e636bac4b\") " pod="openshift-machine-config-operator/machine-config-server-8sbd2" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195130 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/93297fe9-2486-4b24-b6c6-cb917ab240c5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195146 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/93297fe9-2486-4b24-b6c6-cb917ab240c5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195171 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/02fd0443-734e-4aad-ad80-b776e53ded19-etcd-service-ca\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195186 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/980879e9-818d-414c-97c0-ad63e6cf4039-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xl2bw\" (UID: \"980879e9-818d-414c-97c0-ad63e6cf4039\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195219 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195235 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fd0443-734e-4aad-ad80-b776e53ded19-config\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195249 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f1416be-5367-4e50-8d31-3b10657f02df-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2p679\" (UID: \"9f1416be-5367-4e50-8d31-3b10657f02df\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195267 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9157f430-c56e-4b51-bd07-4c7eb99f707c-config-volume\") pod \"dns-default-hbmrd\" (UID: \"9157f430-c56e-4b51-bd07-4c7eb99f707c\") " pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195292 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnf2p\" (UniqueName: \"kubernetes.io/projected/9f1416be-5367-4e50-8d31-3b10657f02df-kube-api-access-rnf2p\") pod \"openshift-apiserver-operator-796bbdcf4f-2p679\" (UID: \"9f1416be-5367-4e50-8d31-3b10657f02df\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195343 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29dz7\" (UniqueName: \"kubernetes.io/projected/9157f430-c56e-4b51-bd07-4c7eb99f707c-kube-api-access-29dz7\") pod \"dns-default-hbmrd\" (UID: \"9157f430-c56e-4b51-bd07-4c7eb99f707c\") " pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195363 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d078a0ae-522e-4694-98e9-8cae5563a71f-srv-cert\") pod \"olm-operator-6b444d44fb-c84rp\" (UID: \"d078a0ae-522e-4694-98e9-8cae5563a71f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195389 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7d716278-6e64-4e98-b111-20228be9d375-etcd-client\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195407 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae4bca1d-744f-45c1-bb86-97713834ff4d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mzdnj\" (UID: \"ae4bca1d-744f-45c1-bb86-97713834ff4d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.195444 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blq49\" (UniqueName: \"kubernetes.io/projected/7d716278-6e64-4e98-b111-20228be9d375-kube-api-access-blq49\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.201938 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae4bca1d-744f-45c1-bb86-97713834ff4d-config\") pod \"kube-apiserver-operator-766d6c64bb-mzdnj\" (UID: \"ae4bca1d-744f-45c1-bb86-97713834ff4d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202219 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zpqp\" (UniqueName: \"kubernetes.io/projected/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-kube-api-access-2zpqp\") pod \"marketplace-operator-79b997595-4zs2z\" (UID: \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202259 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e9a4efdc-7efb-4f90-97f2-ba975c3b172c-srv-cert\") pod \"catalog-operator-68c6474976-v8kgk\" (UID: \"e9a4efdc-7efb-4f90-97f2-ba975c3b172c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202284 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dqwh\" (UniqueName: \"kubernetes.io/projected/f950cada-6ee4-447d-b82f-0cfbdc3e98f7-kube-api-access-2dqwh\") pod \"service-ca-operator-777779d784-6gv5x\" (UID: \"f950cada-6ee4-447d-b82f-0cfbdc3e98f7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202317 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10325f8c-feb6-4dee-8d8c-9ef797f51175-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rqp28\" (UID: \"10325f8c-feb6-4dee-8d8c-9ef797f51175\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202347 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ac15726e-cd7f-4836-b88c-d832e321ba93-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t9jm9\" (UID: \"ac15726e-cd7f-4836-b88c-d832e321ba93\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t9jm9" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202367 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfbvr\" (UniqueName: \"kubernetes.io/projected/ac15726e-cd7f-4836-b88c-d832e321ba93-kube-api-access-vfbvr\") pod \"multus-admission-controller-857f4d67dd-t9jm9\" (UID: \"ac15726e-cd7f-4836-b88c-d832e321ba93\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t9jm9" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202386 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/02fd0443-734e-4aad-ad80-b776e53ded19-etcd-ca\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202404 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b6c59839-1a09-4b0b-90ac-eb9d001e9581-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pgslt\" (UID: \"b6c59839-1a09-4b0b-90ac-eb9d001e9581\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202422 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47377b8d-2f9c-4a58-a99b-a94111193e23-cert\") pod \"ingress-canary-2sv7b\" (UID: \"47377b8d-2f9c-4a58-a99b-a94111193e23\") " pod="openshift-ingress-canary/ingress-canary-2sv7b" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202440 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-registration-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202464 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-socket-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202482 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9157f430-c56e-4b51-bd07-4c7eb99f707c-metrics-tls\") pod \"dns-default-hbmrd\" (UID: \"9157f430-c56e-4b51-bd07-4c7eb99f707c\") " pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202509 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c6dc9fc9-525e-4508-940f-626e636bac4b-certs\") pod \"machine-config-server-8sbd2\" (UID: \"c6dc9fc9-525e-4508-940f-626e636bac4b\") " pod="openshift-machine-config-operator/machine-config-server-8sbd2" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202559 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/92d953e3-3660-4178-a6ee-5c28d37c08fa-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8km27\" (UID: \"92d953e3-3660-4178-a6ee-5c28d37c08fa\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202586 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-etcd-serving-ca\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202603 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10325f8c-feb6-4dee-8d8c-9ef797f51175-proxy-tls\") pod \"machine-config-controller-84d6567774-rqp28\" (UID: \"10325f8c-feb6-4dee-8d8c-9ef797f51175\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.202707 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7d716278-6e64-4e98-b111-20228be9d375-audit-dir\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.203430 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-audit\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: E1129 00:02:17.203480 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:17.703466649 +0000 UTC m=+150.865359881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.203949 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7d716278-6e64-4e98-b111-20228be9d375-node-pullsecrets\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.204010 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c-images\") pod \"machine-config-operator-74547568cd-bfzjb\" (UID: \"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.204162 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-config\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.204503 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f1416be-5367-4e50-8d31-3b10657f02df-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2p679\" (UID: \"9f1416be-5367-4e50-8d31-3b10657f02df\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.205839 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/980879e9-818d-414c-97c0-ad63e6cf4039-config\") pod \"machine-api-operator-5694c8668f-xl2bw\" (UID: \"980879e9-818d-414c-97c0-ad63e6cf4039\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.206009 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10325f8c-feb6-4dee-8d8c-9ef797f51175-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rqp28\" (UID: \"10325f8c-feb6-4dee-8d8c-9ef797f51175\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.206530 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/93297fe9-2486-4b24-b6c6-cb917ab240c5-registry-certificates\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.208072 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/02fd0443-734e-4aad-ad80-b776e53ded19-etcd-client\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.208437 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52e44859-bfe5-4b11-8f91-5722be14dbf7-config\") pod \"kube-controller-manager-operator-78b949d7b-vw8ff\" (UID: \"52e44859-bfe5-4b11-8f91-5722be14dbf7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.209086 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93297fe9-2486-4b24-b6c6-cb917ab240c5-trusted-ca\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.209455 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae4bca1d-744f-45c1-bb86-97713834ff4d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mzdnj\" (UID: \"ae4bca1d-744f-45c1-bb86-97713834ff4d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.209982 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bfzjb\" (UID: \"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.211088 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae4bca1d-744f-45c1-bb86-97713834ff4d-config\") pod \"kube-apiserver-operator-766d6c64bb-mzdnj\" (UID: \"ae4bca1d-744f-45c1-bb86-97713834ff4d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.211262 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-j4wm9"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.212451 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/980879e9-818d-414c-97c0-ad63e6cf4039-images\") pod \"machine-api-operator-5694c8668f-xl2bw\" (UID: \"980879e9-818d-414c-97c0-ad63e6cf4039\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.212761 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52e44859-bfe5-4b11-8f91-5722be14dbf7-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-vw8ff\" (UID: \"52e44859-bfe5-4b11-8f91-5722be14dbf7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.213146 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/02fd0443-734e-4aad-ad80-b776e53ded19-etcd-ca\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.214109 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d078a0ae-522e-4694-98e9-8cae5563a71f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-c84rp\" (UID: \"d078a0ae-522e-4694-98e9-8cae5563a71f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.214162 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/10325f8c-feb6-4dee-8d8c-9ef797f51175-proxy-tls\") pod \"machine-config-controller-84d6567774-rqp28\" (UID: \"10325f8c-feb6-4dee-8d8c-9ef797f51175\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.214507 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02fd0443-734e-4aad-ad80-b776e53ded19-serving-cert\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.215381 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/02fd0443-734e-4aad-ad80-b776e53ded19-etcd-service-ca\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.215596 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fd0443-734e-4aad-ad80-b776e53ded19-config\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.216106 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92d953e3-3660-4178-a6ee-5c28d37c08fa-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8km27\" (UID: \"92d953e3-3660-4178-a6ee-5c28d37c08fa\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.217051 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-image-import-ca\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.217702 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-etcd-serving-ca\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.218314 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d716278-6e64-4e98-b111-20228be9d375-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.222257 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/93297fe9-2486-4b24-b6c6-cb917ab240c5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.223720 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b6c59839-1a09-4b0b-90ac-eb9d001e9581-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pgslt\" (UID: \"b6c59839-1a09-4b0b-90ac-eb9d001e9581\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.223838 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92d953e3-3660-4178-a6ee-5c28d37c08fa-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8km27\" (UID: \"92d953e3-3660-4178-a6ee-5c28d37c08fa\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.225345 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.227752 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/980879e9-818d-414c-97c0-ad63e6cf4039-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xl2bw\" (UID: \"980879e9-818d-414c-97c0-ad63e6cf4039\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.230254 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d078a0ae-522e-4694-98e9-8cae5563a71f-srv-cert\") pod \"olm-operator-6b444d44fb-c84rp\" (UID: \"d078a0ae-522e-4694-98e9-8cae5563a71f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.231074 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7d716278-6e64-4e98-b111-20228be9d375-etcd-client\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.233098 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/93297fe9-2486-4b24-b6c6-cb917ab240c5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.233405 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d716278-6e64-4e98-b111-20228be9d375-serving-cert\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.236945 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-registry-tls\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.237040 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f1416be-5367-4e50-8d31-3b10657f02df-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2p679\" (UID: \"9f1416be-5367-4e50-8d31-3b10657f02df\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.242710 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgkqz\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-kube-api-access-kgkqz\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.243612 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ac15726e-cd7f-4836-b88c-d832e321ba93-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t9jm9\" (UID: \"ac15726e-cd7f-4836-b88c-d832e321ba93\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t9jm9" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.252564 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7d716278-6e64-4e98-b111-20228be9d375-encryption-config\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.256174 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c-proxy-tls\") pod \"machine-config-operator-74547568cd-bfzjb\" (UID: \"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.262160 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-bound-sa-token\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.266246 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.303730 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.303969 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-socket-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304073 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9157f430-c56e-4b51-bd07-4c7eb99f707c-metrics-tls\") pod \"dns-default-hbmrd\" (UID: \"9157f430-c56e-4b51-bd07-4c7eb99f707c\") " pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304189 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c6dc9fc9-525e-4508-940f-626e636bac4b-certs\") pod \"machine-config-server-8sbd2\" (UID: \"c6dc9fc9-525e-4508-940f-626e636bac4b\") " pod="openshift-machine-config-operator/machine-config-server-8sbd2" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304268 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh895\" (UniqueName: \"kubernetes.io/projected/c772f56c-3fa9-4331-b439-ceea5f985452-kube-api-access-zh895\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304343 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4zs2z\" (UID: \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304428 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab-signing-cabundle\") pod \"service-ca-9c57cc56f-7c58s\" (UID: \"dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab\") " pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304489 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-socket-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304557 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f950cada-6ee4-447d-b82f-0cfbdc3e98f7-config\") pod \"service-ca-operator-777779d784-6gv5x\" (UID: \"f950cada-6ee4-447d-b82f-0cfbdc3e98f7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" Nov 29 00:02:17 crc kubenswrapper[4931]: E1129 00:02:17.304643 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:17.804597333 +0000 UTC m=+150.966490565 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304676 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sxqq\" (UniqueName: \"kubernetes.io/projected/c6dc9fc9-525e-4508-940f-626e636bac4b-kube-api-access-2sxqq\") pod \"machine-config-server-8sbd2\" (UID: \"c6dc9fc9-525e-4508-940f-626e636bac4b\") " pod="openshift-machine-config-operator/machine-config-server-8sbd2" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304719 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f950cada-6ee4-447d-b82f-0cfbdc3e98f7-serving-cert\") pod \"service-ca-operator-777779d784-6gv5x\" (UID: \"f950cada-6ee4-447d-b82f-0cfbdc3e98f7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304746 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-csi-data-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304778 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9bd673da-0ab0-468d-adb3-fbc74d2a091a-apiservice-cert\") pod \"packageserver-d55dfcdfc-9xcvz\" (UID: \"9bd673da-0ab0-468d-adb3-fbc74d2a091a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304800 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9bd673da-0ab0-468d-adb3-fbc74d2a091a-tmpfs\") pod \"packageserver-d55dfcdfc-9xcvz\" (UID: \"9bd673da-0ab0-468d-adb3-fbc74d2a091a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304886 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4zs2z\" (UID: \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304908 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fcmd\" (UniqueName: \"kubernetes.io/projected/9bd673da-0ab0-468d-adb3-fbc74d2a091a-kube-api-access-6fcmd\") pod \"packageserver-d55dfcdfc-9xcvz\" (UID: \"9bd673da-0ab0-468d-adb3-fbc74d2a091a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304950 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-plugins-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304995 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305027 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4f7k\" (UniqueName: \"kubernetes.io/projected/dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab-kube-api-access-m4f7k\") pod \"service-ca-9c57cc56f-7c58s\" (UID: \"dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab\") " pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305047 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96bbcdc9-9b57-409e-b5b9-79a99034208a-secret-volume\") pod \"collect-profiles-29406240-fjsn7\" (UID: \"96bbcdc9-9b57-409e-b5b9-79a99034208a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305077 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e9a4efdc-7efb-4f90-97f2-ba975c3b172c-profile-collector-cert\") pod \"catalog-operator-68c6474976-v8kgk\" (UID: \"e9a4efdc-7efb-4f90-97f2-ba975c3b172c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305096 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-mountpoint-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305169 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab-signing-key\") pod \"service-ca-9c57cc56f-7c58s\" (UID: \"dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab\") " pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305191 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ck2r\" (UniqueName: \"kubernetes.io/projected/96bbcdc9-9b57-409e-b5b9-79a99034208a-kube-api-access-6ck2r\") pod \"collect-profiles-29406240-fjsn7\" (UID: \"96bbcdc9-9b57-409e-b5b9-79a99034208a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305214 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96bbcdc9-9b57-409e-b5b9-79a99034208a-config-volume\") pod \"collect-profiles-29406240-fjsn7\" (UID: \"96bbcdc9-9b57-409e-b5b9-79a99034208a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305244 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr2jn\" (UniqueName: \"kubernetes.io/projected/e9a4efdc-7efb-4f90-97f2-ba975c3b172c-kube-api-access-tr2jn\") pod \"catalog-operator-68c6474976-v8kgk\" (UID: \"e9a4efdc-7efb-4f90-97f2-ba975c3b172c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305300 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9bd673da-0ab0-468d-adb3-fbc74d2a091a-webhook-cert\") pod \"packageserver-d55dfcdfc-9xcvz\" (UID: \"9bd673da-0ab0-468d-adb3-fbc74d2a091a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305319 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct58h\" (UniqueName: \"kubernetes.io/projected/47377b8d-2f9c-4a58-a99b-a94111193e23-kube-api-access-ct58h\") pod \"ingress-canary-2sv7b\" (UID: \"47377b8d-2f9c-4a58-a99b-a94111193e23\") " pod="openshift-ingress-canary/ingress-canary-2sv7b" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305341 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c6dc9fc9-525e-4508-940f-626e636bac4b-node-bootstrap-token\") pod \"machine-config-server-8sbd2\" (UID: \"c6dc9fc9-525e-4508-940f-626e636bac4b\") " pod="openshift-machine-config-operator/machine-config-server-8sbd2" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305372 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9157f430-c56e-4b51-bd07-4c7eb99f707c-config-volume\") pod \"dns-default-hbmrd\" (UID: \"9157f430-c56e-4b51-bd07-4c7eb99f707c\") " pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305396 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29dz7\" (UniqueName: \"kubernetes.io/projected/9157f430-c56e-4b51-bd07-4c7eb99f707c-kube-api-access-29dz7\") pod \"dns-default-hbmrd\" (UID: \"9157f430-c56e-4b51-bd07-4c7eb99f707c\") " pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305444 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zpqp\" (UniqueName: \"kubernetes.io/projected/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-kube-api-access-2zpqp\") pod \"marketplace-operator-79b997595-4zs2z\" (UID: \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305465 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e9a4efdc-7efb-4f90-97f2-ba975c3b172c-srv-cert\") pod \"catalog-operator-68c6474976-v8kgk\" (UID: \"e9a4efdc-7efb-4f90-97f2-ba975c3b172c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305495 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dqwh\" (UniqueName: \"kubernetes.io/projected/f950cada-6ee4-447d-b82f-0cfbdc3e98f7-kube-api-access-2dqwh\") pod \"service-ca-operator-777779d784-6gv5x\" (UID: \"f950cada-6ee4-447d-b82f-0cfbdc3e98f7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305535 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47377b8d-2f9c-4a58-a99b-a94111193e23-cert\") pod \"ingress-canary-2sv7b\" (UID: \"47377b8d-2f9c-4a58-a99b-a94111193e23\") " pod="openshift-ingress-canary/ingress-canary-2sv7b" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305551 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-registration-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305641 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-registration-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305955 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.305996 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.306319 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f950cada-6ee4-447d-b82f-0cfbdc3e98f7-config\") pod \"service-ca-operator-777779d784-6gv5x\" (UID: \"f950cada-6ee4-447d-b82f-0cfbdc3e98f7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.308455 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f950cada-6ee4-447d-b82f-0cfbdc3e98f7-serving-cert\") pod \"service-ca-operator-777779d784-6gv5x\" (UID: \"f950cada-6ee4-447d-b82f-0cfbdc3e98f7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.308832 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-csi-data-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.308970 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9157f430-c56e-4b51-bd07-4c7eb99f707c-metrics-tls\") pod \"dns-default-hbmrd\" (UID: \"9157f430-c56e-4b51-bd07-4c7eb99f707c\") " pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.304105 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cb5l\" (UniqueName: \"kubernetes.io/projected/b6c59839-1a09-4b0b-90ac-eb9d001e9581-kube-api-access-9cb5l\") pod \"control-plane-machine-set-operator-78cbb6b69f-pgslt\" (UID: \"b6c59839-1a09-4b0b-90ac-eb9d001e9581\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.312098 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9bd673da-0ab0-468d-adb3-fbc74d2a091a-tmpfs\") pod \"packageserver-d55dfcdfc-9xcvz\" (UID: \"9bd673da-0ab0-468d-adb3-fbc74d2a091a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.312526 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9bd673da-0ab0-468d-adb3-fbc74d2a091a-apiservice-cert\") pod \"packageserver-d55dfcdfc-9xcvz\" (UID: \"9bd673da-0ab0-468d-adb3-fbc74d2a091a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: E1129 00:02:17.312766 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:17.81275341 +0000 UTC m=+150.974646642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.313118 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab-signing-key\") pod \"service-ca-9c57cc56f-7c58s\" (UID: \"dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab\") " pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.313580 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4zs2z\" (UID: \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.313757 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-plugins-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.313787 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9157f430-c56e-4b51-bd07-4c7eb99f707c-config-volume\") pod \"dns-default-hbmrd\" (UID: \"9157f430-c56e-4b51-bd07-4c7eb99f707c\") " pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.314296 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdttl\" (UniqueName: \"kubernetes.io/projected/02fd0443-734e-4aad-ad80-b776e53ded19-kube-api-access-bdttl\") pod \"etcd-operator-b45778765-pr87s\" (UID: \"02fd0443-734e-4aad-ad80-b776e53ded19\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.315209 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c772f56c-3fa9-4331-b439-ceea5f985452-mountpoint-dir\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.316036 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab-signing-cabundle\") pod \"service-ca-9c57cc56f-7c58s\" (UID: \"dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab\") " pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.316605 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96bbcdc9-9b57-409e-b5b9-79a99034208a-config-volume\") pod \"collect-profiles-29406240-fjsn7\" (UID: \"96bbcdc9-9b57-409e-b5b9-79a99034208a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.318403 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9bd673da-0ab0-468d-adb3-fbc74d2a091a-webhook-cert\") pod \"packageserver-d55dfcdfc-9xcvz\" (UID: \"9bd673da-0ab0-468d-adb3-fbc74d2a091a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.319831 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c6dc9fc9-525e-4508-940f-626e636bac4b-node-bootstrap-token\") pod \"machine-config-server-8sbd2\" (UID: \"c6dc9fc9-525e-4508-940f-626e636bac4b\") " pod="openshift-machine-config-operator/machine-config-server-8sbd2" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.325189 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnx8p\" (UniqueName: \"kubernetes.io/projected/233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c-kube-api-access-lnx8p\") pod \"machine-config-operator-74547568cd-bfzjb\" (UID: \"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.325328 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c6dc9fc9-525e-4508-940f-626e636bac4b-certs\") pod \"machine-config-server-8sbd2\" (UID: \"c6dc9fc9-525e-4508-940f-626e636bac4b\") " pod="openshift-machine-config-operator/machine-config-server-8sbd2" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.325715 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96bbcdc9-9b57-409e-b5b9-79a99034208a-secret-volume\") pod \"collect-profiles-29406240-fjsn7\" (UID: \"96bbcdc9-9b57-409e-b5b9-79a99034208a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.325893 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4zs2z\" (UID: \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.326323 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e9a4efdc-7efb-4f90-97f2-ba975c3b172c-srv-cert\") pod \"catalog-operator-68c6474976-v8kgk\" (UID: \"e9a4efdc-7efb-4f90-97f2-ba975c3b172c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.328152 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e9a4efdc-7efb-4f90-97f2-ba975c3b172c-profile-collector-cert\") pod \"catalog-operator-68c6474976-v8kgk\" (UID: \"e9a4efdc-7efb-4f90-97f2-ba975c3b172c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.331289 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29406240-4vs86"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.356666 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.362228 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/47377b8d-2f9c-4a58-a99b-a94111193e23-cert\") pod \"ingress-canary-2sv7b\" (UID: \"47377b8d-2f9c-4a58-a99b-a94111193e23\") " pod="openshift-ingress-canary/ingress-canary-2sv7b" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.401623 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ljv5\" (UniqueName: \"kubernetes.io/projected/d078a0ae-522e-4694-98e9-8cae5563a71f-kube-api-access-7ljv5\") pod \"olm-operator-6b444d44fb-c84rp\" (UID: \"d078a0ae-522e-4694-98e9-8cae5563a71f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.401629 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnf2p\" (UniqueName: \"kubernetes.io/projected/9f1416be-5367-4e50-8d31-3b10657f02df-kube-api-access-rnf2p\") pod \"openshift-apiserver-operator-796bbdcf4f-2p679\" (UID: \"9f1416be-5367-4e50-8d31-3b10657f02df\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.406294 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:17 crc kubenswrapper[4931]: E1129 00:02:17.406692 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:17.906669744 +0000 UTC m=+151.068562976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.411409 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.413292 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blq49\" (UniqueName: \"kubernetes.io/projected/7d716278-6e64-4e98-b111-20228be9d375-kube-api-access-blq49\") pod \"apiserver-76f77b778f-rx5zl\" (UID: \"7d716278-6e64-4e98-b111-20228be9d375\") " pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.413392 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-nlp5n"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.425657 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vvmnt"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.426886 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfbvr\" (UniqueName: \"kubernetes.io/projected/ac15726e-cd7f-4836-b88c-d832e321ba93-kube-api-access-vfbvr\") pod \"multus-admission-controller-857f4d67dd-t9jm9\" (UID: \"ac15726e-cd7f-4836-b88c-d832e321ba93\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t9jm9" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.432222 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.435074 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjgmn\" (UniqueName: \"kubernetes.io/projected/980879e9-818d-414c-97c0-ad63e6cf4039-kube-api-access-kjgmn\") pod \"machine-api-operator-5694c8668f-xl2bw\" (UID: \"980879e9-818d-414c-97c0-ad63e6cf4039\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.444958 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae4bca1d-744f-45c1-bb86-97713834ff4d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mzdnj\" (UID: \"ae4bca1d-744f-45c1-bb86-97713834ff4d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.464398 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.466870 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/92d953e3-3660-4178-a6ee-5c28d37c08fa-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8km27\" (UID: \"92d953e3-3660-4178-a6ee-5c28d37c08fa\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.484578 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52e44859-bfe5-4b11-8f91-5722be14dbf7-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-vw8ff\" (UID: \"52e44859-bfe5-4b11-8f91-5722be14dbf7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.486327 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.495471 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.498278 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.508556 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.508695 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: E1129 00:02:17.509246 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:18.009231569 +0000 UTC m=+151.171124801 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.513017 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-t9jm9" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.517737 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm4bj\" (UniqueName: \"kubernetes.io/projected/10325f8c-feb6-4dee-8d8c-9ef797f51175-kube-api-access-rm4bj\") pod \"machine-config-controller-84d6567774-rqp28\" (UID: \"10325f8c-feb6-4dee-8d8c-9ef797f51175\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.534238 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.537892 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sxqq\" (UniqueName: \"kubernetes.io/projected/c6dc9fc9-525e-4508-940f-626e636bac4b-kube-api-access-2sxqq\") pod \"machine-config-server-8sbd2\" (UID: \"c6dc9fc9-525e-4508-940f-626e636bac4b\") " pod="openshift-machine-config-operator/machine-config-server-8sbd2" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.549139 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.554920 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.565335 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh895\" (UniqueName: \"kubernetes.io/projected/c772f56c-3fa9-4331-b439-ceea5f985452-kube-api-access-zh895\") pod \"csi-hostpathplugin-p4jwd\" (UID: \"c772f56c-3fa9-4331-b439-ceea5f985452\") " pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.590903 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4f7k\" (UniqueName: \"kubernetes.io/projected/dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab-kube-api-access-m4f7k\") pod \"service-ca-9c57cc56f-7c58s\" (UID: \"dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab\") " pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.610772 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9v6hx"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.611499 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:17 crc kubenswrapper[4931]: E1129 00:02:17.611859 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:18.111797025 +0000 UTC m=+151.273690257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.617499 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fcmd\" (UniqueName: \"kubernetes.io/projected/9bd673da-0ab0-468d-adb3-fbc74d2a091a-kube-api-access-6fcmd\") pod \"packageserver-d55dfcdfc-9xcvz\" (UID: \"9bd673da-0ab0-468d-adb3-fbc74d2a091a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.623021 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29dz7\" (UniqueName: \"kubernetes.io/projected/9157f430-c56e-4b51-bd07-4c7eb99f707c-kube-api-access-29dz7\") pod \"dns-default-hbmrd\" (UID: \"9157f430-c56e-4b51-bd07-4c7eb99f707c\") " pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.629370 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8sbd2" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.645660 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zpqp\" (UniqueName: \"kubernetes.io/projected/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-kube-api-access-2zpqp\") pod \"marketplace-operator-79b997595-4zs2z\" (UID: \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.653558 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.686177 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ck2r\" (UniqueName: \"kubernetes.io/projected/96bbcdc9-9b57-409e-b5b9-79a99034208a-kube-api-access-6ck2r\") pod \"collect-profiles-29406240-fjsn7\" (UID: \"96bbcdc9-9b57-409e-b5b9-79a99034208a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.695856 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kshsm"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.697661 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.703872 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ffwsb"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.703918 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.705454 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr2jn\" (UniqueName: \"kubernetes.io/projected/e9a4efdc-7efb-4f90-97f2-ba975c3b172c-kube-api-access-tr2jn\") pod \"catalog-operator-68c6474976-v8kgk\" (UID: \"e9a4efdc-7efb-4f90-97f2-ba975c3b172c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.709038 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.723007 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: E1129 00:02:17.723592 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:18.223560328 +0000 UTC m=+151.385453560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.729425 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dqwh\" (UniqueName: \"kubernetes.io/projected/f950cada-6ee4-447d-b82f-0cfbdc3e98f7-kube-api-access-2dqwh\") pod \"service-ca-operator-777779d784-6gv5x\" (UID: \"f950cada-6ee4-447d-b82f-0cfbdc3e98f7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.730412 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.750258 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct58h\" (UniqueName: \"kubernetes.io/projected/47377b8d-2f9c-4a58-a99b-a94111193e23-kube-api-access-ct58h\") pod \"ingress-canary-2sv7b\" (UID: \"47377b8d-2f9c-4a58-a99b-a94111193e23\") " pod="openshift-ingress-canary/ingress-canary-2sv7b" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.757121 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.769785 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-f2666"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.824689 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:17 crc kubenswrapper[4931]: E1129 00:02:17.824988 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:18.324931559 +0000 UTC m=+151.486824791 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.825221 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:17 crc kubenswrapper[4931]: E1129 00:02:17.825800 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:18.325784973 +0000 UTC m=+151.487678205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:17 crc kubenswrapper[4931]: W1129 00:02:17.857969 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfa259bd_b798_4cd3_9560_10528387a9b2.slice/crio-210e86517d349f10cc3e272442571cb3592cc580c2a84c3659cf8ada1bcd050f WatchSource:0}: Error finding container 210e86517d349f10cc3e272442571cb3592cc580c2a84c3659cf8ada1bcd050f: Status 404 returned error can't find the container with id 210e86517d349f10cc3e272442571cb3592cc580c2a84c3659cf8ada1bcd050f Nov 29 00:02:17 crc kubenswrapper[4931]: W1129 00:02:17.865110 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c5e5b6b_ab0b_4158_9f5a_b6630b901e60.slice/crio-10a6f20ce978a557c25a739ea0e27d5f6b97cf0709dae42ea7e2bcad5c503274 WatchSource:0}: Error finding container 10a6f20ce978a557c25a739ea0e27d5f6b97cf0709dae42ea7e2bcad5c503274: Status 404 returned error can't find the container with id 10a6f20ce978a557c25a739ea0e27d5f6b97cf0709dae42ea7e2bcad5c503274 Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.869915 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.878864 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.885879 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.888631 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rx5zl"] Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.896155 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:17 crc kubenswrapper[4931]: W1129 00:02:17.897835 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod199fba37_4269_4fb2_9249_b384fc2f2905.slice/crio-a50bcd352c4e32faa2853bccdcdb5544622ed59f9ed0bd0e442be657d803c135 WatchSource:0}: Error finding container a50bcd352c4e32faa2853bccdcdb5544622ed59f9ed0bd0e442be657d803c135: Status 404 returned error can't find the container with id a50bcd352c4e32faa2853bccdcdb5544622ed59f9ed0bd0e442be657d803c135 Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.901592 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.907661 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.914444 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.922061 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2sv7b" Nov 29 00:02:17 crc kubenswrapper[4931]: I1129 00:02:17.941205 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:17 crc kubenswrapper[4931]: E1129 00:02:17.941747 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:18.441728498 +0000 UTC m=+151.603621730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.040328 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt"] Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.041570 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679"] Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.046476 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:18 crc kubenswrapper[4931]: E1129 00:02:18.046802 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:18.546789256 +0000 UTC m=+151.708682488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.149214 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:18 crc kubenswrapper[4931]: E1129 00:02:18.149385 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:18.649358882 +0000 UTC m=+151.811252114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.149605 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:18 crc kubenswrapper[4931]: E1129 00:02:18.150174 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:18.650153475 +0000 UTC m=+151.812046707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.203974 4931 generic.go:334] "Generic (PLEG): container finished" podID="695bbf28-25a7-4ba8-a262-fc333dd7707a" containerID="3ed46ccabbc4a934e17c77a10410ecace2ac83ba789c873ac0833e02ae324ae5" exitCode=0 Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.204054 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" event={"ID":"695bbf28-25a7-4ba8-a262-fc333dd7707a","Type":"ContainerDied","Data":"3ed46ccabbc4a934e17c77a10410ecace2ac83ba789c873ac0833e02ae324ae5"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.204080 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" event={"ID":"695bbf28-25a7-4ba8-a262-fc333dd7707a","Type":"ContainerStarted","Data":"d0000d043bb05ba780cb2ad96a239f05e83de6946edf41fc453c79a43b000831"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.207024 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" event={"ID":"127076b5-1262-4859-837f-c1f844eafebf","Type":"ContainerStarted","Data":"5ddd0668f63aef7ef33b982828c062e923221216421465adc58fb56fff144a37"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.216436 4931 generic.go:334] "Generic (PLEG): container finished" podID="c7297960-2485-4860-8587-04ebdf29ad6a" containerID="fea4b46b224658a7a274c546ae0077264f52338d629fd8e8ada36981431b47e1" exitCode=0 Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.216526 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" event={"ID":"c7297960-2485-4860-8587-04ebdf29ad6a","Type":"ContainerDied","Data":"fea4b46b224658a7a274c546ae0077264f52338d629fd8e8ada36981431b47e1"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.236530 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8sbd2" event={"ID":"c6dc9fc9-525e-4508-940f-626e636bac4b","Type":"ContainerStarted","Data":"d603b50c85c86b4308b7b747171d0a422c702eff6aa233855c9e56c71459598d"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.251858 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:18 crc kubenswrapper[4931]: E1129 00:02:18.252299 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:18.752250637 +0000 UTC m=+151.914143869 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.271914 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29406240-4vs86" event={"ID":"35311d49-bd1d-4351-90a8-f332a6c3a5d2","Type":"ContainerStarted","Data":"b1627da317dd8fc82d7909fa64c1b799c974d40852b9e2c0e88f504805d294f4"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.271974 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29406240-4vs86" event={"ID":"35311d49-bd1d-4351-90a8-f332a6c3a5d2","Type":"ContainerStarted","Data":"641c61212b90b06b20e26399aaf0090e858e8f40bd3eb7ec570937f515cb5926"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.276975 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" event={"ID":"bfa259bd-b798-4cd3-9560-10528387a9b2","Type":"ContainerStarted","Data":"210e86517d349f10cc3e272442571cb3592cc580c2a84c3659cf8ada1bcd050f"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.282738 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" event={"ID":"13b49f2e-d3eb-483c-a910-202c955bfda9","Type":"ContainerStarted","Data":"8c3ef93cdb8373c42212ea35ddaae366733ffdc5fafa935c984ce36295945551"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.295606 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" event={"ID":"b40393fc-4539-4caa-a76c-6eda95f7483c","Type":"ContainerStarted","Data":"cd5e9c3a383d64497600f574d5819d45c35d0c78db2ce1f9a8d600268e401c9a"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.295699 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" event={"ID":"b40393fc-4539-4caa-a76c-6eda95f7483c","Type":"ContainerStarted","Data":"aab6d9b15594a80c3683748474780ed5f91ec2bb05df5e2461559e4b411dedc5"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.297870 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" event={"ID":"199fba37-4269-4fb2-9249-b384fc2f2905","Type":"ContainerStarted","Data":"a50bcd352c4e32faa2853bccdcdb5544622ed59f9ed0bd0e442be657d803c135"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.300451 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" event={"ID":"321d4bf6-7bff-4513-87e5-adce01251a0c","Type":"ContainerStarted","Data":"ff5457a984be5b80ffc75da348e45c4796af4c3e5a68be1659df74432f6bbf1c"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.311172 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-plmcw" event={"ID":"bd34298a-7f8c-4264-97d5-4503c5182e49","Type":"ContainerStarted","Data":"09d4ed0c0f3fcc10ce3d5b03e7b9134c00c14cc65adcb72661a06c3be11d33cd"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.314079 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" event={"ID":"0be48302-a154-4099-b4d7-3e4781964c7d","Type":"ContainerStarted","Data":"5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.314124 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" event={"ID":"0be48302-a154-4099-b4d7-3e4781964c7d","Type":"ContainerStarted","Data":"578a78f40900e9f24d3acfbdc6f00519b470506d18c38261eeb0867104620453"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.314453 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.320826 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx" event={"ID":"831e6629-6933-4be3-8d45-a4a9504387d1","Type":"ContainerStarted","Data":"bfb9b3d9cc918a122da1e82e81eb0523168e83eb42d2f8014cce7b9fc2e385b4"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.320875 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx" event={"ID":"831e6629-6933-4be3-8d45-a4a9504387d1","Type":"ContainerStarted","Data":"81a8fe524ed2d28520864b968d5e6587601d835d9ce3c531a4aa48c2cb1abd68"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.324713 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" event={"ID":"7d716278-6e64-4e98-b111-20228be9d375","Type":"ContainerStarted","Data":"7ae3e59819055fd63574d694481b5fd9baffbca4d07989476e7e13b1a686652f"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.325689 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vvmnt" event={"ID":"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864","Type":"ContainerStarted","Data":"405fb4605ca1de717c05ff6fa684bcff2ce130e3dc15a3a0fd5f4e31475d69ff"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.330115 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-j4wm9" event={"ID":"e49dd8f4-f866-4878-b867-d4d64de070c7","Type":"ContainerStarted","Data":"c160cf8831cdcdf7f41425d013aad2945c75f18c6bfee7632824e795098f9000"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.330448 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-j4wm9" event={"ID":"e49dd8f4-f866-4878-b867-d4d64de070c7","Type":"ContainerStarted","Data":"1270f1b421b46ca13b37be29160ff13919b95cac01cfc792bcd1fe1138a24ed3"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.330691 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-j4wm9" Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.333593 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-f2666" event={"ID":"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60","Type":"ContainerStarted","Data":"10a6f20ce978a557c25a739ea0e27d5f6b97cf0709dae42ea7e2bcad5c503274"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.343632 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9v6hx" event={"ID":"0e2c10ac-c727-42af-80a9-6e14937b986c","Type":"ContainerStarted","Data":"7fb3fb169c0e35000628b67f536a42c4f13c458405f4d97e4a0fea5e6db0ea3e"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.351604 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" event={"ID":"2c7ebccc-ceef-4268-8143-5318f56e5337","Type":"ContainerStarted","Data":"dd77eb4880480efd95dad7922a0abd15996ab1eeb63a819a39b5f00a104c3d88"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.353846 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:18 crc kubenswrapper[4931]: E1129 00:02:18.356044 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:18.856027398 +0000 UTC m=+152.017920640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.359231 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-nlp5n" event={"ID":"7ac25448-7ec8-4cb7-8f19-1dab208323a4","Type":"ContainerStarted","Data":"190f3ed72b84118b7275ae36f2dea56db4da23988355e3e199c4e5d9f304095b"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.372714 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-j4wm9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.372776 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-j4wm9" podUID="e49dd8f4-f866-4878-b867-d4d64de070c7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.389904 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" event={"ID":"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72","Type":"ContainerStarted","Data":"8d8632b052d6d059481b0c5fb63b0417260827561b06c7d3a5d697dca3d95ae7"} Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.423882 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pr87s"] Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.453549 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.454882 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:18 crc kubenswrapper[4931]: E1129 00:02:18.456310 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:18.956286906 +0000 UTC m=+152.118180138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.494654 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xl2bw"] Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.506546 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t9jm9"] Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.516119 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27"] Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.525433 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj"] Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.527670 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28"] Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.568933 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:18 crc kubenswrapper[4931]: E1129 00:02:18.572025 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:19.071969183 +0000 UTC m=+152.233862415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.608765 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.710971 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:18 crc kubenswrapper[4931]: E1129 00:02:18.711648 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:19.211620028 +0000 UTC m=+152.373513260 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.739679 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-lfh9z" podStartSLOduration=126.739659884 podStartE2EDuration="2m6.739659884s" podCreationTimestamp="2025-11-29 00:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:18.70447177 +0000 UTC m=+151.866365012" watchObservedRunningTime="2025-11-29 00:02:18.739659884 +0000 UTC m=+151.901553116" Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.742479 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-p4jwd"] Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.774456 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb"] Nov 29 00:02:18 crc kubenswrapper[4931]: W1129 00:02:18.783294 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae4bca1d_744f_45c1_bb86_97713834ff4d.slice/crio-da3c5da2ab0f780b70bbde383107e131166bd0d46a97c6a9d0c3428e96db8220 WatchSource:0}: Error finding container da3c5da2ab0f780b70bbde383107e131166bd0d46a97c6a9d0c3428e96db8220: Status 404 returned error can't find the container with id da3c5da2ab0f780b70bbde383107e131166bd0d46a97c6a9d0c3428e96db8220 Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.789320 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp"] Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.814625 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:18 crc kubenswrapper[4931]: E1129 00:02:18.814940 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:19.314927725 +0000 UTC m=+152.476820957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.916703 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:18 crc kubenswrapper[4931]: E1129 00:02:18.917024 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:19.417006107 +0000 UTC m=+152.578899339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.922427 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.930382 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:18 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:18 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:18 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:18 crc kubenswrapper[4931]: I1129 00:02:18.930427 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.006675 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff"] Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.019968 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:19.519956153 +0000 UTC m=+152.681849385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.020031 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.137539 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.137861 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:19.637848425 +0000 UTC m=+152.799741657 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.240059 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.241753 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:19.741738189 +0000 UTC m=+152.903631421 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.341044 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.341490 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:19.841463492 +0000 UTC m=+153.003356724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.341633 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.341921 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:19.841915055 +0000 UTC m=+153.003808287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.377180 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-vvmnt" podStartSLOduration=126.377163971 podStartE2EDuration="2m6.377163971s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:19.37678326 +0000 UTC m=+152.538676492" watchObservedRunningTime="2025-11-29 00:02:19.377163971 +0000 UTC m=+152.539057203" Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.402183 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t9jm9" event={"ID":"ac15726e-cd7f-4836-b88c-d832e321ba93","Type":"ContainerStarted","Data":"f4a44a5f19b730f7816cef9cfad2ae90ef6ac538e2c7a6b59884153742f3414c"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.406338 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9v6hx" event={"ID":"0e2c10ac-c727-42af-80a9-6e14937b986c","Type":"ContainerStarted","Data":"5c94ee407479247eb1615601058f7763f9b6a1be1214c369530ed7b0abf84398"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.413491 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" event={"ID":"13b49f2e-d3eb-483c-a910-202c955bfda9","Type":"ContainerStarted","Data":"73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.414950 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.427974 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" event={"ID":"9f1416be-5367-4e50-8d31-3b10657f02df","Type":"ContainerStarted","Data":"44d3379da439811b3e727b6355f6f02f7a00e12112a0dc04568f7d4c903cbfeb"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.434013 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-f2666" event={"ID":"3c5e5b6b-ab0b-4158-9f5a-b6630b901e60","Type":"ContainerStarted","Data":"47377c06e3ee26fdf2975cd6f8cd2e088158afce79d8c271fd1532df495740a7"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.434574 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.439336 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.442170 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.442524 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:19.942509483 +0000 UTC m=+153.104402705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.442623 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.442934 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:19.942926855 +0000 UTC m=+153.104820087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.457025 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt" event={"ID":"b6c59839-1a09-4b0b-90ac-eb9d001e9581","Type":"ContainerStarted","Data":"1034e441332804d2451a0ec1fe2358fc2da6236df383aaad438e780e0528416b"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.475926 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" event={"ID":"02fd0443-734e-4aad-ad80-b776e53ded19","Type":"ContainerStarted","Data":"f7cabcf9a9ae71a1e179ff2c6e2d94b6f2c6a253e10106a6e419c9e52b928a7b"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.480791 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" event={"ID":"d078a0ae-522e-4694-98e9-8cae5563a71f","Type":"ContainerStarted","Data":"8519c54b73a5d581cf29cc7edc77424494e09a5375cdd9db86d57c077b296624"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.481737 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz"] Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.525237 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx" event={"ID":"831e6629-6933-4be3-8d45-a4a9504387d1","Type":"ContainerStarted","Data":"ff60ef1d43de95dbc7f0fa93182aef8688c337a598adf521c91748c6e729db40"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.543204 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.544070 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:20.044054659 +0000 UTC m=+153.205947891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.571243 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-nlp5n" event={"ID":"7ac25448-7ec8-4cb7-8f19-1dab208323a4","Type":"ContainerStarted","Data":"5f789eb0f4b80e4cc267e10d032d39136d4389f37da5644db8fa6e816d5f769e"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.581302 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" event={"ID":"f6ac7072-d5d5-4c7a-8ea1-89090be1fb72","Type":"ContainerStarted","Data":"d25636a73a47213ae54fcb2a6bc5a0fa7cefcebee7e2796c9204143cfb818ffc"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.623163 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8sbd2" event={"ID":"c6dc9fc9-525e-4508-940f-626e636bac4b","Type":"ContainerStarted","Data":"2d59ad743a7e046155db79762bdc0019565fa92f93eba45aa44b3f49b6bfc8cd"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.631699 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" podStartSLOduration=126.631674239 podStartE2EDuration="2m6.631674239s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:19.623061038 +0000 UTC m=+152.784954280" watchObservedRunningTime="2025-11-29 00:02:19.631674239 +0000 UTC m=+152.793567481" Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.641569 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" event={"ID":"ae4bca1d-744f-45c1-bb86-97713834ff4d","Type":"ContainerStarted","Data":"da3c5da2ab0f780b70bbde383107e131166bd0d46a97c6a9d0c3428e96db8220"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.645551 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.645879 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:20.145867412 +0000 UTC m=+153.307760644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.654479 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" event={"ID":"980879e9-818d-414c-97c0-ad63e6cf4039","Type":"ContainerStarted","Data":"914c74dda7ca1ab5ab115ef64c54c9d19a414a65d07632d48c3ca18e02e46878"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.682865 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x"] Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.683701 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7c58s"] Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.705539 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2sv7b"] Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.707221 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" event={"ID":"bfa259bd-b798-4cd3-9560-10528387a9b2","Type":"ContainerStarted","Data":"e34d1e80a4d5cd094a80fb8452d11ecea0a33c305a521ed2efb11be21b6504c3"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.742693 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hzbr7" podStartSLOduration=126.74267541 podStartE2EDuration="2m6.74267541s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:19.720174795 +0000 UTC m=+152.882068047" watchObservedRunningTime="2025-11-29 00:02:19.74267541 +0000 UTC m=+152.904568642" Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.743078 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7"] Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.743112 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vvmnt" event={"ID":"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864","Type":"ContainerStarted","Data":"160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.746707 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.747651 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:20.247634824 +0000 UTC m=+153.409528056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.751663 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk"] Nov 29 00:02:19 crc kubenswrapper[4931]: W1129 00:02:19.755784 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfe24ce5_b8e4_4e7d_9801_f1df0601e0ab.slice/crio-7c406fe101e98d4a64f92b2dfb30495e0e8586d94fd658a75e53ee0fdd965912 WatchSource:0}: Error finding container 7c406fe101e98d4a64f92b2dfb30495e0e8586d94fd658a75e53ee0fdd965912: Status 404 returned error can't find the container with id 7c406fe101e98d4a64f92b2dfb30495e0e8586d94fd658a75e53ee0fdd965912 Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.756635 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.783900 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lpxjw" podStartSLOduration=126.783878419 podStartE2EDuration="2m6.783878419s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:19.760850819 +0000 UTC m=+152.922744061" watchObservedRunningTime="2025-11-29 00:02:19.783878419 +0000 UTC m=+152.945771651" Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.784668 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" event={"ID":"c772f56c-3fa9-4331-b439-ceea5f985452","Type":"ContainerStarted","Data":"803a8563b5716ee6f54c20cb9b3732cf5abf037a464e4377457ef258418fffed"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.800044 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" event={"ID":"10325f8c-feb6-4dee-8d8c-9ef797f51175","Type":"ContainerStarted","Data":"2e01ea8f745f2c0e2d0e312c4728be28bca5ecb5e8b815c8e669594bffcbb0e6"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.841777 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hbmrd"] Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.849765 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.850361 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" event={"ID":"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c","Type":"ContainerStarted","Data":"550f02e2b2c0b70944c728c44f90f3ea610e97168c5a6cd9e83c09f23d77a4f6"} Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.851484 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:20.351472017 +0000 UTC m=+153.513365249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.862494 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" event={"ID":"52e44859-bfe5-4b11-8f91-5722be14dbf7","Type":"ContainerStarted","Data":"b1a7c8f57fd9da4105b7d1dda8f1e1d61d94075e71be6129dbe0a281882bb118"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.871895 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" event={"ID":"92d953e3-3660-4178-a6ee-5c28d37c08fa","Type":"ContainerStarted","Data":"e201f661d3c8c412ac0961aefba787e9e4bfc47340ccc1ca9065e8cf84b71a75"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.882729 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4zs2z"] Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.930650 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:19 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:19 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:19 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.930693 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.932414 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" event={"ID":"321d4bf6-7bff-4513-87e5-adce01251a0c","Type":"ContainerStarted","Data":"7100cb767a5ec2cf932cdbdea0ca924d79cbbf99b623f93746f2b96c75b3cd46"} Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.932495 4931 patch_prober.go:28] interesting pod/downloads-7954f5f757-j4wm9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.932527 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-j4wm9" podUID="e49dd8f4-f866-4878-b867-d4d64de070c7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Nov 29 00:02:19 crc kubenswrapper[4931]: W1129 00:02:19.948225 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62bb1b45_d5f5_4a0d_8c8d_212acc620a3a.slice/crio-7e61c3011cf88f7cab6824f1504f670559e1a17e78910d7175bb15249f5c1e03 WatchSource:0}: Error finding container 7e61c3011cf88f7cab6824f1504f670559e1a17e78910d7175bb15249f5c1e03: Status 404 returned error can't find the container with id 7e61c3011cf88f7cab6824f1504f670559e1a17e78910d7175bb15249f5c1e03 Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.951178 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.951284 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:20.451264052 +0000 UTC m=+153.613157294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.951433 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:19 crc kubenswrapper[4931]: E1129 00:02:19.951773 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:20.451758476 +0000 UTC m=+153.613651708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:19 crc kubenswrapper[4931]: I1129 00:02:19.961536 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29406240-4vs86" podStartSLOduration=126.96151718 podStartE2EDuration="2m6.96151718s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:19.904655375 +0000 UTC m=+153.066548607" watchObservedRunningTime="2025-11-29 00:02:19.96151718 +0000 UTC m=+153.123410412" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.033359 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-j4wm9" podStartSLOduration=127.03334369 podStartE2EDuration="2m7.03334369s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:19.9900704 +0000 UTC m=+153.151963642" watchObservedRunningTime="2025-11-29 00:02:20.03334369 +0000 UTC m=+153.195236922" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.055687 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:20 crc kubenswrapper[4931]: E1129 00:02:20.057605 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:20.557584375 +0000 UTC m=+153.719477607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.077456 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-plmcw" podStartSLOduration=127.077440993 podStartE2EDuration="2m7.077440993s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:20.076552027 +0000 UTC m=+153.238445259" watchObservedRunningTime="2025-11-29 00:02:20.077440993 +0000 UTC m=+153.239334225" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.101017 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-f2666" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.146776 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-h25zx" podStartSLOduration=127.146762671 podStartE2EDuration="2m7.146762671s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:20.146065301 +0000 UTC m=+153.307958533" watchObservedRunningTime="2025-11-29 00:02:20.146762671 +0000 UTC m=+153.308655903" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.158092 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:20 crc kubenswrapper[4931]: E1129 00:02:20.158371 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:20.658351488 +0000 UTC m=+153.820244720 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.176732 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" podStartSLOduration=127.176718683 podStartE2EDuration="2m7.176718683s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:20.176143436 +0000 UTC m=+153.338036688" watchObservedRunningTime="2025-11-29 00:02:20.176718683 +0000 UTC m=+153.338611915" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.234688 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" podStartSLOduration=127.23467385 podStartE2EDuration="2m7.23467385s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:20.207946102 +0000 UTC m=+153.369839344" watchObservedRunningTime="2025-11-29 00:02:20.23467385 +0000 UTC m=+153.396567082" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.235013 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-8sbd2" podStartSLOduration=6.23500815 podStartE2EDuration="6.23500815s" podCreationTimestamp="2025-11-29 00:02:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:20.232294481 +0000 UTC m=+153.394187723" watchObservedRunningTime="2025-11-29 00:02:20.23500815 +0000 UTC m=+153.396901382" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.259916 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:20 crc kubenswrapper[4931]: E1129 00:02:20.260197 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:20.760184423 +0000 UTC m=+153.922077655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.310933 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9c86t" podStartSLOduration=128.310918989 podStartE2EDuration="2m8.310918989s" podCreationTimestamp="2025-11-29 00:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:20.284235163 +0000 UTC m=+153.446128385" watchObservedRunningTime="2025-11-29 00:02:20.310918989 +0000 UTC m=+153.472812221" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.311392 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-f2666" podStartSLOduration=127.311386803 podStartE2EDuration="2m7.311386803s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:20.309349844 +0000 UTC m=+153.471243086" watchObservedRunningTime="2025-11-29 00:02:20.311386803 +0000 UTC m=+153.473280035" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.362548 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:20 crc kubenswrapper[4931]: E1129 00:02:20.362799 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:20.862787689 +0000 UTC m=+154.024680921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.403238 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" podStartSLOduration=127.403221136 podStartE2EDuration="2m7.403221136s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:20.362157681 +0000 UTC m=+153.524050923" watchObservedRunningTime="2025-11-29 00:02:20.403221136 +0000 UTC m=+153.565114358" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.466510 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:20 crc kubenswrapper[4931]: E1129 00:02:20.466764 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:20.966749455 +0000 UTC m=+154.128642677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.580714 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:20 crc kubenswrapper[4931]: E1129 00:02:20.581046 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:21.081034682 +0000 UTC m=+154.242927914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.681445 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:20 crc kubenswrapper[4931]: E1129 00:02:20.681844 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:21.181828896 +0000 UTC m=+154.343722128 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.756891 4931 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-kshsm container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.8:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.757276 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.8:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.782892 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:20 crc kubenswrapper[4931]: E1129 00:02:20.783224 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:21.283211307 +0000 UTC m=+154.445104549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.888889 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:20 crc kubenswrapper[4931]: E1129 00:02:20.889153 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:21.38913769 +0000 UTC m=+154.551030922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.926977 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:20 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:20 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:20 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.927266 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.964536 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" event={"ID":"199fba37-4269-4fb2-9249-b384fc2f2905","Type":"ContainerStarted","Data":"dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f"} Nov 29 00:02:20 crc kubenswrapper[4931]: I1129 00:02:20.990726 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:20 crc kubenswrapper[4931]: E1129 00:02:20.991058 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:21.491046737 +0000 UTC m=+154.652939969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.003166 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" event={"ID":"dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab","Type":"ContainerStarted","Data":"7c406fe101e98d4a64f92b2dfb30495e0e8586d94fd658a75e53ee0fdd965912"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.020131 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hbmrd" event={"ID":"9157f430-c56e-4b51-bd07-4c7eb99f707c","Type":"ContainerStarted","Data":"369b7564032311f59c4fff18a8942b4bd556ae3c6be1995d6e0df566594b05b5"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.027931 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" event={"ID":"bfa259bd-b798-4cd3-9560-10528387a9b2","Type":"ContainerStarted","Data":"ee2d185356e71f5b042a162b63f143ae379bea1768dd46376f8f1348797e3740"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.028843 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.029945 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" event={"ID":"f950cada-6ee4-447d-b82f-0cfbdc3e98f7","Type":"ContainerStarted","Data":"19b84ee291210778a097a5d863077f25d130decbaf48498bc962b28280068a88"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.029975 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" event={"ID":"f950cada-6ee4-447d-b82f-0cfbdc3e98f7","Type":"ContainerStarted","Data":"1f87d0714a8083daedf265e74c4af8ea1624b34aba450f1d47527d16a88c15b8"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.040418 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" event={"ID":"9bd673da-0ab0-468d-adb3-fbc74d2a091a","Type":"ContainerStarted","Data":"e7d5d60ecc90787af372aa6f7d120c72f5da879d5185159e4102815dfeca0b45"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.040461 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" event={"ID":"9bd673da-0ab0-468d-adb3-fbc74d2a091a","Type":"ContainerStarted","Data":"aeec02cebf9bb04a3e8b6ee876d8faf33398bbeccae27d25b7cc016441e30ce8"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.041267 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.045307 4931 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-9xcvz container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.045347 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" podUID="9bd673da-0ab0-468d-adb3-fbc74d2a091a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.078699 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" podStartSLOduration=128.078684007 podStartE2EDuration="2m8.078684007s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.076072671 +0000 UTC m=+154.237965913" watchObservedRunningTime="2025-11-29 00:02:21.078684007 +0000 UTC m=+154.240577239" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.080934 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" event={"ID":"52e44859-bfe5-4b11-8f91-5722be14dbf7","Type":"ContainerStarted","Data":"61c796eb2c50919ec38086803d5b8e8dcb15d0c14600a201a09645096ed2953f"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.091398 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:21 crc kubenswrapper[4931]: E1129 00:02:21.092580 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:21.592564291 +0000 UTC m=+154.754457523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.108898 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-njwv4" event={"ID":"2c7ebccc-ceef-4268-8143-5318f56e5337","Type":"ContainerStarted","Data":"73b85197b11d1ed0d4bc762104e5d201e2171a0862bb3018857adb5165d758e7"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.122502 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" event={"ID":"e9a4efdc-7efb-4f90-97f2-ba975c3b172c","Type":"ContainerStarted","Data":"d62331564a824f0a84691e2e1341d0f154e7d67134b1dd7a075f8fb5f73b58ac"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.123484 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.126383 4931 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-v8kgk container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.126418 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" podUID="e9a4efdc-7efb-4f90-97f2-ba975c3b172c" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.126569 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" event={"ID":"ae4bca1d-744f-45c1-bb86-97713834ff4d","Type":"ContainerStarted","Data":"06aaf9e28d035e674559856de8b222db7aab995ae79d84ab21c98d71cb17e2b3"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.134787 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-nlp5n" event={"ID":"7ac25448-7ec8-4cb7-8f19-1dab208323a4","Type":"ContainerStarted","Data":"32f28807e7e79877c6529525110a52a4275f0205c07fda311c62840241557a6a"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.139685 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6gv5x" podStartSLOduration=128.139674013 podStartE2EDuration="2m8.139674013s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.109732711 +0000 UTC m=+154.271625943" watchObservedRunningTime="2025-11-29 00:02:21.139674013 +0000 UTC m=+154.301567245" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.140254 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" podStartSLOduration=128.140249539 podStartE2EDuration="2m8.140249539s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.138698984 +0000 UTC m=+154.300592226" watchObservedRunningTime="2025-11-29 00:02:21.140249539 +0000 UTC m=+154.302142761" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.168876 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" event={"ID":"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c","Type":"ContainerStarted","Data":"d3e4db4b7f1096855aa8ffc76be32fabd334e5de7c6d11ce9e1442f46a0f710c"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.168922 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" event={"ID":"233e88cc-1e3f-42a2-9327-e6fcd2ab1d9c","Type":"ContainerStarted","Data":"5a2893a2a7466ad8d6a06c258502b2108f2d72a6420e93c6c96984d216a58828"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.169152 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzdnj" podStartSLOduration=128.16913479 podStartE2EDuration="2m8.16913479s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.167348658 +0000 UTC m=+154.329241900" watchObservedRunningTime="2025-11-29 00:02:21.16913479 +0000 UTC m=+154.331028022" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.187735 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9v6hx" event={"ID":"0e2c10ac-c727-42af-80a9-6e14937b986c","Type":"ContainerStarted","Data":"c3d363da25efce685bc629041a7bfd46b6eb6e93a6aef35ccb2745aaed3b3f61"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.205593 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:21 crc kubenswrapper[4931]: E1129 00:02:21.208061 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:21.708048833 +0000 UTC m=+154.869942065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.236140 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" event={"ID":"92d953e3-3660-4178-a6ee-5c28d37c08fa","Type":"ContainerStarted","Data":"aeadf2bade0c041e0c5701d5f638a5461de94fdff5da720d8fd0bebb86e5f9ed"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.240717 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2sv7b" event={"ID":"47377b8d-2f9c-4a58-a99b-a94111193e23","Type":"ContainerStarted","Data":"0e312a49a97974911da0af11db7e4b98b320f3fa8d7cb16703a89621b27171dc"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.240780 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2sv7b" event={"ID":"47377b8d-2f9c-4a58-a99b-a94111193e23","Type":"ContainerStarted","Data":"3eb44f1cda1b54a695eba9fd2634103f769fcdb2e6b34816fbe4dd1d904f371b"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.262222 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" podStartSLOduration=128.262205719 podStartE2EDuration="2m8.262205719s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.214373027 +0000 UTC m=+154.376266269" watchObservedRunningTime="2025-11-29 00:02:21.262205719 +0000 UTC m=+154.424098951" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.268154 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" event={"ID":"d078a0ae-522e-4694-98e9-8cae5563a71f","Type":"ContainerStarted","Data":"4e386b0571a7507011b23fae137df59a1f916b1edbbc1e312561760a9069a089"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.268905 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.275138 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-nlp5n" podStartSLOduration=128.275115715 podStartE2EDuration="2m8.275115715s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.261336564 +0000 UTC m=+154.423229806" watchObservedRunningTime="2025-11-29 00:02:21.275115715 +0000 UTC m=+154.437008947" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.298079 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt" event={"ID":"b6c59839-1a09-4b0b-90ac-eb9d001e9581","Type":"ContainerStarted","Data":"bc50169945fdf03b645db2595077c29af5329267cc1eaeff2d523f915d4134eb"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.311292 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vw8ff" podStartSLOduration=128.311276008 podStartE2EDuration="2m8.311276008s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.309156926 +0000 UTC m=+154.471050158" watchObservedRunningTime="2025-11-29 00:02:21.311276008 +0000 UTC m=+154.473169240" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.313703 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:21 crc kubenswrapper[4931]: E1129 00:02:21.314764 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:21.814742749 +0000 UTC m=+154.976635991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.328339 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" event={"ID":"96bbcdc9-9b57-409e-b5b9-79a99034208a","Type":"ContainerStarted","Data":"4da6124bf83dca21ab93286bffc7edaf51f2333742a5d124c8169da9cd027cb0"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.329095 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.342781 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9v6hx" podStartSLOduration=128.342764684 podStartE2EDuration="2m8.342764684s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.341684463 +0000 UTC m=+154.503577695" watchObservedRunningTime="2025-11-29 00:02:21.342764684 +0000 UTC m=+154.504657916" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.361155 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.370997 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t9jm9" event={"ID":"ac15726e-cd7f-4836-b88c-d832e321ba93","Type":"ContainerStarted","Data":"d5d7a66aea813173f08b14aa2b972c9cc9d7aff4bc272dbbf7c18960cc1fd7a5"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.386038 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" event={"ID":"127076b5-1262-4859-837f-c1f844eafebf","Type":"ContainerStarted","Data":"4c1d38fe25507d3825d45423bab7d008e1695d4ed81a9714eb2621be42f9be39"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.386082 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" event={"ID":"127076b5-1262-4859-837f-c1f844eafebf","Type":"ContainerStarted","Data":"f3beccd1cfd6a9be8d809c307b403eed369a04bca93d3ee7c4776820e6843266"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.402712 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" event={"ID":"c772f56c-3fa9-4331-b439-ceea5f985452","Type":"ContainerStarted","Data":"4e436750cff9638dc94e72ce1f7d7f4a27f5968e0e93010956e0b38536f8c36b"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.415337 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:21 crc kubenswrapper[4931]: E1129 00:02:21.415745 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:21.915729798 +0000 UTC m=+155.077623020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.420524 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" event={"ID":"02fd0443-734e-4aad-ad80-b776e53ded19","Type":"ContainerStarted","Data":"b27a23012bba3a3485ed7c02b21f5f4ce0cd629556d7eee4846bb88a46576eb2"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.426676 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" event={"ID":"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a","Type":"ContainerStarted","Data":"7e61c3011cf88f7cab6824f1504f670559e1a17e78910d7175bb15249f5c1e03"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.427518 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.427735 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bfzjb" podStartSLOduration=128.427720537 podStartE2EDuration="2m8.427720537s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.391884904 +0000 UTC m=+154.553778136" watchObservedRunningTime="2025-11-29 00:02:21.427720537 +0000 UTC m=+154.589613769" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.427913 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-c84rp" podStartSLOduration=128.427908023 podStartE2EDuration="2m8.427908023s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.426462841 +0000 UTC m=+154.588356083" watchObservedRunningTime="2025-11-29 00:02:21.427908023 +0000 UTC m=+154.589801265" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.437441 4931 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4zs2z container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.437508 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" podUID="62bb1b45-d5f5-4a0d-8c8d-212acc620a3a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.469693 4931 generic.go:334] "Generic (PLEG): container finished" podID="7d716278-6e64-4e98-b111-20228be9d375" containerID="587eb580256c26a59502d5db8ce7577b7a98fb43280552aeec228e433d661b50" exitCode=0 Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.469760 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" event={"ID":"7d716278-6e64-4e98-b111-20228be9d375","Type":"ContainerDied","Data":"587eb580256c26a59502d5db8ce7577b7a98fb43280552aeec228e433d661b50"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.493183 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" event={"ID":"695bbf28-25a7-4ba8-a262-fc333dd7707a","Type":"ContainerStarted","Data":"9ece687af0c00556f6b18d60815ad48714e47c95f6d45d12145f3814823f68a8"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.494014 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.505747 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8km27" podStartSLOduration=128.505731638 podStartE2EDuration="2m8.505731638s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.455696611 +0000 UTC m=+154.617589843" watchObservedRunningTime="2025-11-29 00:02:21.505731638 +0000 UTC m=+154.667624870" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.518244 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:21 crc kubenswrapper[4931]: E1129 00:02:21.519538 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:22.019517089 +0000 UTC m=+155.181410321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.535870 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-2sv7b" podStartSLOduration=7.535805343 podStartE2EDuration="7.535805343s" podCreationTimestamp="2025-11-29 00:02:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.534546476 +0000 UTC m=+154.696439728" watchObservedRunningTime="2025-11-29 00:02:21.535805343 +0000 UTC m=+154.697698575" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.536517 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pgslt" podStartSLOduration=128.536513404 podStartE2EDuration="2m8.536513404s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.506384777 +0000 UTC m=+154.668278019" watchObservedRunningTime="2025-11-29 00:02:21.536513404 +0000 UTC m=+154.698406636" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.577062 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" event={"ID":"10325f8c-feb6-4dee-8d8c-9ef797f51175","Type":"ContainerStarted","Data":"2e92fb1a8d2a65a62aaf30223e47cd4b8d44063274acb92b842ee6ff6e6f9045"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.622715 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:21 crc kubenswrapper[4931]: E1129 00:02:21.625412 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:22.125400851 +0000 UTC m=+155.287294083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.637140 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" podStartSLOduration=128.637123432 podStartE2EDuration="2m8.637123432s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.634177167 +0000 UTC m=+154.796070399" watchObservedRunningTime="2025-11-29 00:02:21.637123432 +0000 UTC m=+154.799016654" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.645177 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" event={"ID":"9f1416be-5367-4e50-8d31-3b10657f02df","Type":"ContainerStarted","Data":"c08a18e6a521fcc8134f553e1523a7c131e4d00224354de260657001bfc68f76"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.676603 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" event={"ID":"980879e9-818d-414c-97c0-ad63e6cf4039","Type":"ContainerStarted","Data":"a13e3ccb03dafbd7ace21438f6123c2c3a2ae9843ed914603f9c49d963aa84f2"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.676928 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" event={"ID":"980879e9-818d-414c-97c0-ad63e6cf4039","Type":"ContainerStarted","Data":"c2cdd5188655952f4b104a632031d9453a9f956fd163f884bcc7e616650ef6a3"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.678373 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-t9jm9" podStartSLOduration=128.678356733 podStartE2EDuration="2m8.678356733s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.677333443 +0000 UTC m=+154.839226675" watchObservedRunningTime="2025-11-29 00:02:21.678356733 +0000 UTC m=+154.840249965" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.714972 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-bs2jq" podStartSLOduration=128.714957058 podStartE2EDuration="2m8.714957058s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.714268608 +0000 UTC m=+154.876161850" watchObservedRunningTime="2025-11-29 00:02:21.714957058 +0000 UTC m=+154.876850290" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.715544 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" event={"ID":"c7297960-2485-4860-8587-04ebdf29ad6a","Type":"ContainerStarted","Data":"6adca0964f5f932eea66a42b53a79ab8ac26d3e39d36004ff7853934e563306f"} Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.732396 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:21 crc kubenswrapper[4931]: E1129 00:02:21.733229 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:22.233204909 +0000 UTC m=+155.395098141 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.835590 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:21 crc kubenswrapper[4931]: E1129 00:02:21.844524 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:22.344508029 +0000 UTC m=+155.506401251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.875761 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-pr87s" podStartSLOduration=128.875745878 podStartE2EDuration="2m8.875745878s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.772094221 +0000 UTC m=+154.933987463" watchObservedRunningTime="2025-11-29 00:02:21.875745878 +0000 UTC m=+155.037639100" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.876601 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" podStartSLOduration=128.876595093 podStartE2EDuration="2m8.876595093s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:21.865260353 +0000 UTC m=+155.027153585" watchObservedRunningTime="2025-11-29 00:02:21.876595093 +0000 UTC m=+155.038488325" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.938999 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:21 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:21 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:21 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.939052 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:21 crc kubenswrapper[4931]: I1129 00:02:21.939396 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:21 crc kubenswrapper[4931]: E1129 00:02:21.939707 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:22.4396932 +0000 UTC m=+155.601586432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.040493 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:22 crc kubenswrapper[4931]: E1129 00:02:22.040787 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:22.540775052 +0000 UTC m=+155.702668284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.047590 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" podStartSLOduration=129.04757582 podStartE2EDuration="2m9.04757582s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:22.045543561 +0000 UTC m=+155.207436803" watchObservedRunningTime="2025-11-29 00:02:22.04757582 +0000 UTC m=+155.209469052" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.146316 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:22 crc kubenswrapper[4931]: E1129 00:02:22.146629 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:22.646615293 +0000 UTC m=+155.808508525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.194576 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" podStartSLOduration=129.194559808 podStartE2EDuration="2m9.194559808s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:22.189064658 +0000 UTC m=+155.350957920" watchObservedRunningTime="2025-11-29 00:02:22.194559808 +0000 UTC m=+155.356453040" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.249429 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:22 crc kubenswrapper[4931]: E1129 00:02:22.249762 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:22.749750875 +0000 UTC m=+155.911644107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.295069 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" podStartSLOduration=129.295053273 podStartE2EDuration="2m9.295053273s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:22.24616189 +0000 UTC m=+155.408055122" watchObservedRunningTime="2025-11-29 00:02:22.295053273 +0000 UTC m=+155.456946505" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.337180 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-xl2bw" podStartSLOduration=129.337166839 podStartE2EDuration="2m9.337166839s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:22.296055912 +0000 UTC m=+155.457949154" watchObservedRunningTime="2025-11-29 00:02:22.337166839 +0000 UTC m=+155.499060071" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.350384 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:22 crc kubenswrapper[4931]: E1129 00:02:22.350869 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:22.850854728 +0000 UTC m=+156.012747960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.455314 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:22 crc kubenswrapper[4931]: E1129 00:02:22.455733 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:22.95571407 +0000 UTC m=+156.117607302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.556283 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:22 crc kubenswrapper[4931]: E1129 00:02:22.556440 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.056416031 +0000 UTC m=+156.218309263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.556495 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:22 crc kubenswrapper[4931]: E1129 00:02:22.556849 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.056838503 +0000 UTC m=+156.218731735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.657684 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:22 crc kubenswrapper[4931]: E1129 00:02:22.658059 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.158043669 +0000 UTC m=+156.319936901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.755830 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vbd6v" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.758798 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:22 crc kubenswrapper[4931]: E1129 00:02:22.759111 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.259098021 +0000 UTC m=+156.420991253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.761465 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hbmrd" event={"ID":"9157f430-c56e-4b51-bd07-4c7eb99f707c","Type":"ContainerStarted","Data":"3cf263f4876e35831b687bf41bfe0ca112472cd0ed7c33c4cb0668ec82a46453"} Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.761497 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hbmrd" event={"ID":"9157f430-c56e-4b51-bd07-4c7eb99f707c","Type":"ContainerStarted","Data":"d482cfd74450b3c21803c5412b0efc7fe41e5b6b423485fd04d8a8cd8e4c72ad"} Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.761601 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.774043 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rqp28" event={"ID":"10325f8c-feb6-4dee-8d8c-9ef797f51175","Type":"ContainerStarted","Data":"47e046d23a4e2f2b22c80260a00f74f53ce5f02d4640dd93e21de16451943526"} Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.791641 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" event={"ID":"e9a4efdc-7efb-4f90-97f2-ba975c3b172c","Type":"ContainerStarted","Data":"3910698146d94274f1f25cf3bc4f0f207b52d2b271b53fb6c617a9672952d01d"} Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.798334 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2p679" podStartSLOduration=130.798319792 podStartE2EDuration="2m10.798319792s" podCreationTimestamp="2025-11-29 00:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:22.337448937 +0000 UTC m=+155.499342169" watchObservedRunningTime="2025-11-29 00:02:22.798319792 +0000 UTC m=+155.960213024" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.803232 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v8kgk" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.806983 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" event={"ID":"dfe24ce5-b8e4-4e7d-9801-f1df0601e0ab","Type":"ContainerStarted","Data":"89be0cd0772626b44da5fe8102af8b14e4767f3033bb27c6f5120769536b56b1"} Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.831167 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" event={"ID":"7d716278-6e64-4e98-b111-20228be9d375","Type":"ContainerStarted","Data":"d35baacba3fd7018d1a7d79cb8671788eeffa8ee570947da87cd934d6ebfd325"} Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.831210 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" event={"ID":"7d716278-6e64-4e98-b111-20228be9d375","Type":"ContainerStarted","Data":"0d725705cbdbd207fc3bb2429e3bf6455c059f23bf78cf88359b778762fd7fef"} Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.848775 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" event={"ID":"96bbcdc9-9b57-409e-b5b9-79a99034208a","Type":"ContainerStarted","Data":"a0faded80efe8587290a77ed570ea80cd569c557fbf016a8c3089016d33ab90b"} Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.850695 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-hbmrd" podStartSLOduration=8.850684787 podStartE2EDuration="8.850684787s" podCreationTimestamp="2025-11-29 00:02:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:22.815972966 +0000 UTC m=+155.977866208" watchObservedRunningTime="2025-11-29 00:02:22.850684787 +0000 UTC m=+156.012578019" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.850941 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-7c58s" podStartSLOduration=129.850938394 podStartE2EDuration="2m9.850938394s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:22.849647766 +0000 UTC m=+156.011540998" watchObservedRunningTime="2025-11-29 00:02:22.850938394 +0000 UTC m=+156.012831616" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.859449 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:22 crc kubenswrapper[4931]: E1129 00:02:22.860945 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.360927965 +0000 UTC m=+156.522821187 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.877839 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t9jm9" event={"ID":"ac15726e-cd7f-4836-b88c-d832e321ba93","Type":"ContainerStarted","Data":"3ef45724f6e9dd3d2894e807f0dd5980b02eec545fbc864e345e9d1e0a07337c"} Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.895614 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" event={"ID":"c772f56c-3fa9-4331-b439-ceea5f985452","Type":"ContainerStarted","Data":"2d264f48a10a79c9fbc07096698ecafa744cf213129584a9544a9e7e2c2913d8"} Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.897736 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" event={"ID":"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a","Type":"ContainerStarted","Data":"f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876"} Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.906273 4931 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4zs2z container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.906318 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" podUID="62bb1b45-d5f5-4a0d-8c8d-212acc620a3a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.920253 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" podStartSLOduration=130.920235841 podStartE2EDuration="2m10.920235841s" podCreationTimestamp="2025-11-29 00:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:22.920086277 +0000 UTC m=+156.081979529" watchObservedRunningTime="2025-11-29 00:02:22.920235841 +0000 UTC m=+156.082129063" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.924589 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:22 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:22 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:22 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.924668 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:22 crc kubenswrapper[4931]: I1129 00:02:22.966645 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:22 crc kubenswrapper[4931]: E1129 00:02:22.975986 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.475963453 +0000 UTC m=+156.637856685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.055113 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9xcvz" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.067326 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:23 crc kubenswrapper[4931]: E1129 00:02:23.068490 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.568475136 +0000 UTC m=+156.730368368 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.172306 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:23 crc kubenswrapper[4931]: E1129 00:02:23.172677 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.672662759 +0000 UTC m=+156.834555991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.273834 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:23 crc kubenswrapper[4931]: E1129 00:02:23.274028 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.774003408 +0000 UTC m=+156.935896630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.274355 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:23 crc kubenswrapper[4931]: E1129 00:02:23.274635 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.774623687 +0000 UTC m=+156.936516919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.289219 4931 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.375321 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:23 crc kubenswrapper[4931]: E1129 00:02:23.375516 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.875483672 +0000 UTC m=+157.037376904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.375567 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:23 crc kubenswrapper[4931]: E1129 00:02:23.375869 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.875855983 +0000 UTC m=+157.037749215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.477164 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:23 crc kubenswrapper[4931]: E1129 00:02:23.477360 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.977334387 +0000 UTC m=+157.139227619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.477550 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:23 crc kubenswrapper[4931]: E1129 00:02:23.477877 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-29 00:02:23.977863932 +0000 UTC m=+157.139757164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8j97r" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.502358 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hzhd2"] Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.503245 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.505323 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.520445 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hzhd2"] Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.579022 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.579300 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-catalog-content\") pod \"certified-operators-hzhd2\" (UID: \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\") " pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.579347 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fq6z\" (UniqueName: \"kubernetes.io/projected/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-kube-api-access-6fq6z\") pod \"certified-operators-hzhd2\" (UID: \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\") " pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.579363 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-utilities\") pod \"certified-operators-hzhd2\" (UID: \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\") " pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:02:23 crc kubenswrapper[4931]: E1129 00:02:23.579469 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-29 00:02:24.079453228 +0000 UTC m=+157.241346460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.649200 4931 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-29T00:02:23.289248452Z","Handler":null,"Name":""} Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.674553 4931 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.674593 4931 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.680183 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-catalog-content\") pod \"certified-operators-hzhd2\" (UID: \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\") " pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.680238 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.680259 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fq6z\" (UniqueName: \"kubernetes.io/projected/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-kube-api-access-6fq6z\") pod \"certified-operators-hzhd2\" (UID: \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\") " pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.680275 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-utilities\") pod \"certified-operators-hzhd2\" (UID: \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\") " pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.681049 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-utilities\") pod \"certified-operators-hzhd2\" (UID: \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\") " pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.681375 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-catalog-content\") pod \"certified-operators-hzhd2\" (UID: \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\") " pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.686160 4931 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.686193 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.701382 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fq6z\" (UniqueName: \"kubernetes.io/projected/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-kube-api-access-6fq6z\") pod \"certified-operators-hzhd2\" (UID: \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\") " pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.703866 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r5h8c"] Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.704971 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.707174 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.724313 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8j97r\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.734905 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.735079 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.754456 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r5h8c"] Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.771126 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.781389 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.782045 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39557253-7dac-4846-a945-44b03824ec08-utilities\") pod \"community-operators-r5h8c\" (UID: \"39557253-7dac-4846-a945-44b03824ec08\") " pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.782454 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f65v\" (UniqueName: \"kubernetes.io/projected/39557253-7dac-4846-a945-44b03824ec08-kube-api-access-2f65v\") pod \"community-operators-r5h8c\" (UID: \"39557253-7dac-4846-a945-44b03824ec08\") " pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.782575 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39557253-7dac-4846-a945-44b03824ec08-catalog-content\") pod \"community-operators-r5h8c\" (UID: \"39557253-7dac-4846-a945-44b03824ec08\") " pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.788077 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.816833 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.887528 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39557253-7dac-4846-a945-44b03824ec08-utilities\") pod \"community-operators-r5h8c\" (UID: \"39557253-7dac-4846-a945-44b03824ec08\") " pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.887578 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f65v\" (UniqueName: \"kubernetes.io/projected/39557253-7dac-4846-a945-44b03824ec08-kube-api-access-2f65v\") pod \"community-operators-r5h8c\" (UID: \"39557253-7dac-4846-a945-44b03824ec08\") " pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.887621 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39557253-7dac-4846-a945-44b03824ec08-catalog-content\") pod \"community-operators-r5h8c\" (UID: \"39557253-7dac-4846-a945-44b03824ec08\") " pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.888267 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39557253-7dac-4846-a945-44b03824ec08-catalog-content\") pod \"community-operators-r5h8c\" (UID: \"39557253-7dac-4846-a945-44b03824ec08\") " pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.888928 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39557253-7dac-4846-a945-44b03824ec08-utilities\") pod \"community-operators-r5h8c\" (UID: \"39557253-7dac-4846-a945-44b03824ec08\") " pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.894521 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cdzck"] Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.895452 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.905449 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cdzck"] Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.910100 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f65v\" (UniqueName: \"kubernetes.io/projected/39557253-7dac-4846-a945-44b03824ec08-kube-api-access-2f65v\") pod \"community-operators-r5h8c\" (UID: \"39557253-7dac-4846-a945-44b03824ec08\") " pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.927024 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" event={"ID":"c772f56c-3fa9-4331-b439-ceea5f985452","Type":"ContainerStarted","Data":"cd76e36828225403c5264ca25480a80afb2ebae2ed5963cd9b95c09a9f034ab2"} Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.927056 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" event={"ID":"c772f56c-3fa9-4331-b439-ceea5f985452","Type":"ContainerStarted","Data":"e087dd276a3d8da45197aa623ba72f14f6a5ba6bbbb29eb1e6525dd59a4f0e62"} Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.930134 4931 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4zs2z container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.930168 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" podUID="62bb1b45-d5f5-4a0d-8c8d-212acc620a3a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.931745 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:23 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:23 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:23 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.931783 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.946642 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-p4jwd" podStartSLOduration=9.946624596 podStartE2EDuration="9.946624596s" podCreationTimestamp="2025-11-29 00:02:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:23.945142843 +0000 UTC m=+157.107036085" watchObservedRunningTime="2025-11-29 00:02:23.946624596 +0000 UTC m=+157.108517828" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.992526 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-catalog-content\") pod \"certified-operators-cdzck\" (UID: \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\") " pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.992612 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d92sp\" (UniqueName: \"kubernetes.io/projected/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-kube-api-access-d92sp\") pod \"certified-operators-cdzck\" (UID: \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\") " pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:02:23 crc kubenswrapper[4931]: I1129 00:02:23.992795 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-utilities\") pod \"certified-operators-cdzck\" (UID: \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\") " pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.044674 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.048416 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8j97r"] Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.093790 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-catalog-content\") pod \"certified-operators-cdzck\" (UID: \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\") " pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.093859 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d92sp\" (UniqueName: \"kubernetes.io/projected/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-kube-api-access-d92sp\") pod \"certified-operators-cdzck\" (UID: \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\") " pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.093920 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-utilities\") pod \"certified-operators-cdzck\" (UID: \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\") " pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.094412 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-utilities\") pod \"certified-operators-cdzck\" (UID: \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\") " pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.095413 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-catalog-content\") pod \"certified-operators-cdzck\" (UID: \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\") " pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.098740 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6ngbj"] Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.099734 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.111148 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6ngbj"] Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.124507 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d92sp\" (UniqueName: \"kubernetes.io/projected/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-kube-api-access-d92sp\") pod \"certified-operators-cdzck\" (UID: \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\") " pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.142679 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hzhd2"] Nov 29 00:02:24 crc kubenswrapper[4931]: W1129 00:02:24.160505 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c5af5e4_0d88_443e_8f17_93ce08cb4bab.slice/crio-1b05cbbae8781a790627c92e494b3b5b0edac37d067f55a3ac9fd12aee1cf87b WatchSource:0}: Error finding container 1b05cbbae8781a790627c92e494b3b5b0edac37d067f55a3ac9fd12aee1cf87b: Status 404 returned error can't find the container with id 1b05cbbae8781a790627c92e494b3b5b0edac37d067f55a3ac9fd12aee1cf87b Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.195640 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26p65\" (UniqueName: \"kubernetes.io/projected/bf2c190f-062c-407a-87dd-48a5c4ac07d5-kube-api-access-26p65\") pod \"community-operators-6ngbj\" (UID: \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\") " pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.196509 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf2c190f-062c-407a-87dd-48a5c4ac07d5-catalog-content\") pod \"community-operators-6ngbj\" (UID: \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\") " pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.196592 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf2c190f-062c-407a-87dd-48a5c4ac07d5-utilities\") pod \"community-operators-6ngbj\" (UID: \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\") " pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.220074 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.297396 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26p65\" (UniqueName: \"kubernetes.io/projected/bf2c190f-062c-407a-87dd-48a5c4ac07d5-kube-api-access-26p65\") pod \"community-operators-6ngbj\" (UID: \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\") " pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.297463 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf2c190f-062c-407a-87dd-48a5c4ac07d5-catalog-content\") pod \"community-operators-6ngbj\" (UID: \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\") " pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.297484 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf2c190f-062c-407a-87dd-48a5c4ac07d5-utilities\") pod \"community-operators-6ngbj\" (UID: \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\") " pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.297918 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf2c190f-062c-407a-87dd-48a5c4ac07d5-utilities\") pod \"community-operators-6ngbj\" (UID: \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\") " pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.298195 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf2c190f-062c-407a-87dd-48a5c4ac07d5-catalog-content\") pod \"community-operators-6ngbj\" (UID: \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\") " pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.312907 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26p65\" (UniqueName: \"kubernetes.io/projected/bf2c190f-062c-407a-87dd-48a5c4ac07d5-kube-api-access-26p65\") pod \"community-operators-6ngbj\" (UID: \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\") " pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.428156 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.534109 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r5h8c"] Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.634112 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cdzck"] Nov 29 00:02:24 crc kubenswrapper[4931]: W1129 00:02:24.639374 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e56db6c_4dd6_461b_8f37_5ce71ea285b3.slice/crio-9b06a9337ef5b7ec20ec48dd14448c6e3769b1a25eaf3797923ea2fa352e8810 WatchSource:0}: Error finding container 9b06a9337ef5b7ec20ec48dd14448c6e3769b1a25eaf3797923ea2fa352e8810: Status 404 returned error can't find the container with id 9b06a9337ef5b7ec20ec48dd14448c6e3769b1a25eaf3797923ea2fa352e8810 Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.816122 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6ngbj"] Nov 29 00:02:24 crc kubenswrapper[4931]: W1129 00:02:24.824384 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf2c190f_062c_407a_87dd_48a5c4ac07d5.slice/crio-da718e56cb1d2a7bc362178f791db351ef8e2b3f25f9b976a99fb47b6ceb096f WatchSource:0}: Error finding container da718e56cb1d2a7bc362178f791db351ef8e2b3f25f9b976a99fb47b6ceb096f: Status 404 returned error can't find the container with id da718e56cb1d2a7bc362178f791db351ef8e2b3f25f9b976a99fb47b6ceb096f Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.923902 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:24 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:24 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:24 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.923963 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.934265 4931 generic.go:334] "Generic (PLEG): container finished" podID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" containerID="b5d2161d19f8b120d4d8c56a2c734b8b4bd1cfb7df591debf08c953a1378fd44" exitCode=0 Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.934313 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzhd2" event={"ID":"7c5af5e4-0d88-443e-8f17-93ce08cb4bab","Type":"ContainerDied","Data":"b5d2161d19f8b120d4d8c56a2c734b8b4bd1cfb7df591debf08c953a1378fd44"} Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.934351 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzhd2" event={"ID":"7c5af5e4-0d88-443e-8f17-93ce08cb4bab","Type":"ContainerStarted","Data":"1b05cbbae8781a790627c92e494b3b5b0edac37d067f55a3ac9fd12aee1cf87b"} Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.936387 4931 generic.go:334] "Generic (PLEG): container finished" podID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" containerID="e986bdc9e9e692b98b22b797ae3c873f22b4cab965633438f596d3bea113c0d9" exitCode=0 Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.936472 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cdzck" event={"ID":"3e56db6c-4dd6-461b-8f37-5ce71ea285b3","Type":"ContainerDied","Data":"e986bdc9e9e692b98b22b797ae3c873f22b4cab965633438f596d3bea113c0d9"} Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.936501 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cdzck" event={"ID":"3e56db6c-4dd6-461b-8f37-5ce71ea285b3","Type":"ContainerStarted","Data":"9b06a9337ef5b7ec20ec48dd14448c6e3769b1a25eaf3797923ea2fa352e8810"} Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.938455 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.942191 4931 generic.go:334] "Generic (PLEG): container finished" podID="39557253-7dac-4846-a945-44b03824ec08" containerID="9d6789c2fc8f197dbc1f43572bd0f93b8fa624e96c1f0fd244cbaee3aaf9af63" exitCode=0 Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.942311 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5h8c" event={"ID":"39557253-7dac-4846-a945-44b03824ec08","Type":"ContainerDied","Data":"9d6789c2fc8f197dbc1f43572bd0f93b8fa624e96c1f0fd244cbaee3aaf9af63"} Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.942339 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5h8c" event={"ID":"39557253-7dac-4846-a945-44b03824ec08","Type":"ContainerStarted","Data":"fcb95e718e5ac685f08eca6b7746cc1ff3e8632687e106739395060f11bb1ca6"} Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.943871 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ngbj" event={"ID":"bf2c190f-062c-407a-87dd-48a5c4ac07d5","Type":"ContainerStarted","Data":"da718e56cb1d2a7bc362178f791db351ef8e2b3f25f9b976a99fb47b6ceb096f"} Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.953306 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" event={"ID":"93297fe9-2486-4b24-b6c6-cb917ab240c5","Type":"ContainerStarted","Data":"2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219"} Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.953373 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" event={"ID":"93297fe9-2486-4b24-b6c6-cb917ab240c5","Type":"ContainerStarted","Data":"30f7807fcfda3e2816451654584376de3a71d6980df08546de1998da9ae3de0c"} Nov 29 00:02:24 crc kubenswrapper[4931]: I1129 00:02:24.957834 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.048663 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" podStartSLOduration=132.048635303 podStartE2EDuration="2m12.048635303s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:25.044220225 +0000 UTC m=+158.206113477" watchObservedRunningTime="2025-11-29 00:02:25.048635303 +0000 UTC m=+158.210528615" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.220647 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.495333 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bk72v"] Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.496325 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.499232 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.563354 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk72v"] Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.612419 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c7b21f-8584-482c-b18b-074549e1a120-utilities\") pod \"redhat-marketplace-bk72v\" (UID: \"97c7b21f-8584-482c-b18b-074549e1a120\") " pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.612569 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqf6h\" (UniqueName: \"kubernetes.io/projected/97c7b21f-8584-482c-b18b-074549e1a120-kube-api-access-pqf6h\") pod \"redhat-marketplace-bk72v\" (UID: \"97c7b21f-8584-482c-b18b-074549e1a120\") " pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.612788 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c7b21f-8584-482c-b18b-074549e1a120-catalog-content\") pod \"redhat-marketplace-bk72v\" (UID: \"97c7b21f-8584-482c-b18b-074549e1a120\") " pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.713860 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c7b21f-8584-482c-b18b-074549e1a120-catalog-content\") pod \"redhat-marketplace-bk72v\" (UID: \"97c7b21f-8584-482c-b18b-074549e1a120\") " pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.713925 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c7b21f-8584-482c-b18b-074549e1a120-utilities\") pod \"redhat-marketplace-bk72v\" (UID: \"97c7b21f-8584-482c-b18b-074549e1a120\") " pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.713959 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqf6h\" (UniqueName: \"kubernetes.io/projected/97c7b21f-8584-482c-b18b-074549e1a120-kube-api-access-pqf6h\") pod \"redhat-marketplace-bk72v\" (UID: \"97c7b21f-8584-482c-b18b-074549e1a120\") " pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.714318 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c7b21f-8584-482c-b18b-074549e1a120-catalog-content\") pod \"redhat-marketplace-bk72v\" (UID: \"97c7b21f-8584-482c-b18b-074549e1a120\") " pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.714929 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c7b21f-8584-482c-b18b-074549e1a120-utilities\") pod \"redhat-marketplace-bk72v\" (UID: \"97c7b21f-8584-482c-b18b-074549e1a120\") " pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.743173 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqf6h\" (UniqueName: \"kubernetes.io/projected/97c7b21f-8584-482c-b18b-074549e1a120-kube-api-access-pqf6h\") pod \"redhat-marketplace-bk72v\" (UID: \"97c7b21f-8584-482c-b18b-074549e1a120\") " pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.814512 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.906917 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p2tsz"] Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.908597 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.909199 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2tsz"] Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.929694 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:25 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:25 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:25 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.929734 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.983832 4931 generic.go:334] "Generic (PLEG): container finished" podID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" containerID="3e00e2c6606ce8b1d26fb8ffa2ad0ba3ba2759475341049405ed4e67a5df3538" exitCode=0 Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.984014 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ngbj" event={"ID":"bf2c190f-062c-407a-87dd-48a5c4ac07d5","Type":"ContainerDied","Data":"3e00e2c6606ce8b1d26fb8ffa2ad0ba3ba2759475341049405ed4e67a5df3538"} Nov 29 00:02:25 crc kubenswrapper[4931]: I1129 00:02:25.984943 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.020070 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5b8171-a997-49fe-ac00-eb90258f747a-utilities\") pod \"redhat-marketplace-p2tsz\" (UID: \"2f5b8171-a997-49fe-ac00-eb90258f747a\") " pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.020109 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2s8w\" (UniqueName: \"kubernetes.io/projected/2f5b8171-a997-49fe-ac00-eb90258f747a-kube-api-access-n2s8w\") pod \"redhat-marketplace-p2tsz\" (UID: \"2f5b8171-a997-49fe-ac00-eb90258f747a\") " pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.020224 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5b8171-a997-49fe-ac00-eb90258f747a-catalog-content\") pod \"redhat-marketplace-p2tsz\" (UID: \"2f5b8171-a997-49fe-ac00-eb90258f747a\") " pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.098585 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk72v"] Nov 29 00:02:26 crc kubenswrapper[4931]: W1129 00:02:26.116784 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97c7b21f_8584_482c_b18b_074549e1a120.slice/crio-304d4632cdbe4bebd8359321539d47d8f14a710d1516d64137a6e7803fbd5ca6 WatchSource:0}: Error finding container 304d4632cdbe4bebd8359321539d47d8f14a710d1516d64137a6e7803fbd5ca6: Status 404 returned error can't find the container with id 304d4632cdbe4bebd8359321539d47d8f14a710d1516d64137a6e7803fbd5ca6 Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.121098 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5b8171-a997-49fe-ac00-eb90258f747a-catalog-content\") pod \"redhat-marketplace-p2tsz\" (UID: \"2f5b8171-a997-49fe-ac00-eb90258f747a\") " pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.121233 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5b8171-a997-49fe-ac00-eb90258f747a-utilities\") pod \"redhat-marketplace-p2tsz\" (UID: \"2f5b8171-a997-49fe-ac00-eb90258f747a\") " pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.121252 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2s8w\" (UniqueName: \"kubernetes.io/projected/2f5b8171-a997-49fe-ac00-eb90258f747a-kube-api-access-n2s8w\") pod \"redhat-marketplace-p2tsz\" (UID: \"2f5b8171-a997-49fe-ac00-eb90258f747a\") " pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.121570 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5b8171-a997-49fe-ac00-eb90258f747a-catalog-content\") pod \"redhat-marketplace-p2tsz\" (UID: \"2f5b8171-a997-49fe-ac00-eb90258f747a\") " pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.121909 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5b8171-a997-49fe-ac00-eb90258f747a-utilities\") pod \"redhat-marketplace-p2tsz\" (UID: \"2f5b8171-a997-49fe-ac00-eb90258f747a\") " pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.137581 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2s8w\" (UniqueName: \"kubernetes.io/projected/2f5b8171-a997-49fe-ac00-eb90258f747a-kube-api-access-n2s8w\") pod \"redhat-marketplace-p2tsz\" (UID: \"2f5b8171-a997-49fe-ac00-eb90258f747a\") " pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.238084 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.585738 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.586110 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.592497 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.643153 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2tsz"] Nov 29 00:02:26 crc kubenswrapper[4931]: W1129 00:02:26.659994 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f5b8171_a997_49fe_ac00_eb90258f747a.slice/crio-92ff5b34990be4d69c11fd4e786142bb24760a30af18ab9bf67e760c30b6f654 WatchSource:0}: Error finding container 92ff5b34990be4d69c11fd4e786142bb24760a30af18ab9bf67e760c30b6f654: Status 404 returned error can't find the container with id 92ff5b34990be4d69c11fd4e786142bb24760a30af18ab9bf67e760c30b6f654 Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.693955 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q9c9x"] Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.694907 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.698971 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.703729 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q9c9x"] Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.714715 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-j4wm9" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.732660 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e2b7be6-d4e7-4e92-b419-bda83df88970-catalog-content\") pod \"redhat-operators-q9c9x\" (UID: \"2e2b7be6-d4e7-4e92-b419-bda83df88970\") " pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.732748 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6f2s\" (UniqueName: \"kubernetes.io/projected/2e2b7be6-d4e7-4e92-b419-bda83df88970-kube-api-access-p6f2s\") pod \"redhat-operators-q9c9x\" (UID: \"2e2b7be6-d4e7-4e92-b419-bda83df88970\") " pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.732852 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e2b7be6-d4e7-4e92-b419-bda83df88970-utilities\") pod \"redhat-operators-q9c9x\" (UID: \"2e2b7be6-d4e7-4e92-b419-bda83df88970\") " pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.795549 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.795589 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.799021 4931 patch_prober.go:28] interesting pod/console-f9d7485db-vvmnt container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.799106 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vvmnt" podUID="2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.833726 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6f2s\" (UniqueName: \"kubernetes.io/projected/2e2b7be6-d4e7-4e92-b419-bda83df88970-kube-api-access-p6f2s\") pod \"redhat-operators-q9c9x\" (UID: \"2e2b7be6-d4e7-4e92-b419-bda83df88970\") " pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.833858 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e2b7be6-d4e7-4e92-b419-bda83df88970-utilities\") pod \"redhat-operators-q9c9x\" (UID: \"2e2b7be6-d4e7-4e92-b419-bda83df88970\") " pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.833911 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e2b7be6-d4e7-4e92-b419-bda83df88970-catalog-content\") pod \"redhat-operators-q9c9x\" (UID: \"2e2b7be6-d4e7-4e92-b419-bda83df88970\") " pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.834640 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e2b7be6-d4e7-4e92-b419-bda83df88970-utilities\") pod \"redhat-operators-q9c9x\" (UID: \"2e2b7be6-d4e7-4e92-b419-bda83df88970\") " pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.834827 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e2b7be6-d4e7-4e92-b419-bda83df88970-catalog-content\") pod \"redhat-operators-q9c9x\" (UID: \"2e2b7be6-d4e7-4e92-b419-bda83df88970\") " pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.850691 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6f2s\" (UniqueName: \"kubernetes.io/projected/2e2b7be6-d4e7-4e92-b419-bda83df88970-kube-api-access-p6f2s\") pod \"redhat-operators-q9c9x\" (UID: \"2e2b7be6-d4e7-4e92-b419-bda83df88970\") " pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.921138 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.925260 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:26 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:26 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:26 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.925332 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.965735 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.966587 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.968957 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.969255 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 29 00:02:26 crc kubenswrapper[4931]: I1129 00:02:26.969431 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.004964 4931 generic.go:334] "Generic (PLEG): container finished" podID="2f5b8171-a997-49fe-ac00-eb90258f747a" containerID="3d76e961881be63e5fb0b7aef98a8043913431da1a8852ed7504ffb2720e613e" exitCode=0 Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.005006 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2tsz" event={"ID":"2f5b8171-a997-49fe-ac00-eb90258f747a","Type":"ContainerDied","Data":"3d76e961881be63e5fb0b7aef98a8043913431da1a8852ed7504ffb2720e613e"} Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.005055 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2tsz" event={"ID":"2f5b8171-a997-49fe-ac00-eb90258f747a","Type":"ContainerStarted","Data":"92ff5b34990be4d69c11fd4e786142bb24760a30af18ab9bf67e760c30b6f654"} Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.011973 4931 generic.go:334] "Generic (PLEG): container finished" podID="97c7b21f-8584-482c-b18b-074549e1a120" containerID="5a4d7fb7165df19636715d713472577e37de60837cc92eadbe2112b024524b25" exitCode=0 Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.012034 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk72v" event={"ID":"97c7b21f-8584-482c-b18b-074549e1a120","Type":"ContainerDied","Data":"5a4d7fb7165df19636715d713472577e37de60837cc92eadbe2112b024524b25"} Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.012092 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk72v" event={"ID":"97c7b21f-8584-482c-b18b-074549e1a120","Type":"ContainerStarted","Data":"304d4632cdbe4bebd8359321539d47d8f14a710d1516d64137a6e7803fbd5ca6"} Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.018634 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ncrdv" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.021492 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.037079 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2046696b-1636-47ba-9b22-d5eff643436c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2046696b-1636-47ba-9b22-d5eff643436c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.037355 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2046696b-1636-47ba-9b22-d5eff643436c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2046696b-1636-47ba-9b22-d5eff643436c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.129015 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g72fb"] Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.130045 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.140594 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2046696b-1636-47ba-9b22-d5eff643436c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2046696b-1636-47ba-9b22-d5eff643436c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.140837 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2046696b-1636-47ba-9b22-d5eff643436c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2046696b-1636-47ba-9b22-d5eff643436c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.141010 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2046696b-1636-47ba-9b22-d5eff643436c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2046696b-1636-47ba-9b22-d5eff643436c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.171500 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2046696b-1636-47ba-9b22-d5eff643436c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2046696b-1636-47ba-9b22-d5eff643436c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.211014 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g72fb"] Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.242477 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74999bec-b01b-4643-9cad-fa5e146089f5-catalog-content\") pod \"redhat-operators-g72fb\" (UID: \"74999bec-b01b-4643-9cad-fa5e146089f5\") " pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.242564 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74999bec-b01b-4643-9cad-fa5e146089f5-utilities\") pod \"redhat-operators-g72fb\" (UID: \"74999bec-b01b-4643-9cad-fa5e146089f5\") " pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.242640 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vmjf\" (UniqueName: \"kubernetes.io/projected/74999bec-b01b-4643-9cad-fa5e146089f5-kube-api-access-5vmjf\") pod \"redhat-operators-g72fb\" (UID: \"74999bec-b01b-4643-9cad-fa5e146089f5\") " pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.283767 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.344095 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74999bec-b01b-4643-9cad-fa5e146089f5-catalog-content\") pod \"redhat-operators-g72fb\" (UID: \"74999bec-b01b-4643-9cad-fa5e146089f5\") " pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.344177 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74999bec-b01b-4643-9cad-fa5e146089f5-utilities\") pod \"redhat-operators-g72fb\" (UID: \"74999bec-b01b-4643-9cad-fa5e146089f5\") " pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.344241 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vmjf\" (UniqueName: \"kubernetes.io/projected/74999bec-b01b-4643-9cad-fa5e146089f5-kube-api-access-5vmjf\") pod \"redhat-operators-g72fb\" (UID: \"74999bec-b01b-4643-9cad-fa5e146089f5\") " pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.345860 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74999bec-b01b-4643-9cad-fa5e146089f5-catalog-content\") pod \"redhat-operators-g72fb\" (UID: \"74999bec-b01b-4643-9cad-fa5e146089f5\") " pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.346120 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74999bec-b01b-4643-9cad-fa5e146089f5-utilities\") pod \"redhat-operators-g72fb\" (UID: \"74999bec-b01b-4643-9cad-fa5e146089f5\") " pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.377422 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vmjf\" (UniqueName: \"kubernetes.io/projected/74999bec-b01b-4643-9cad-fa5e146089f5-kube-api-access-5vmjf\") pod \"redhat-operators-g72fb\" (UID: \"74999bec-b01b-4643-9cad-fa5e146089f5\") " pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.465617 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.465687 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.475679 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.482015 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.660742 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q9c9x"] Nov 29 00:02:27 crc kubenswrapper[4931]: W1129 00:02:27.666964 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e2b7be6_d4e7_4e92_b419_bda83df88970.slice/crio-cec7b0fa99c7095c785da0e358bc1d27ad2dfa3c30cd2ff81167ba8543e53ce3 WatchSource:0}: Error finding container cec7b0fa99c7095c785da0e358bc1d27ad2dfa3c30cd2ff81167ba8543e53ce3: Status 404 returned error can't find the container with id cec7b0fa99c7095c785da0e358bc1d27ad2dfa3c30cd2ff81167ba8543e53ce3 Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.740438 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 29 00:02:27 crc kubenswrapper[4931]: W1129 00:02:27.748204 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2046696b_1636_47ba_9b22_d5eff643436c.slice/crio-13a319e8ea3e358ee4fb0475cbac8aa449efe3a9a1fc6d65bfe6e4ed4f8fb2b5 WatchSource:0}: Error finding container 13a319e8ea3e358ee4fb0475cbac8aa449efe3a9a1fc6d65bfe6e4ed4f8fb2b5: Status 404 returned error can't find the container with id 13a319e8ea3e358ee4fb0475cbac8aa449efe3a9a1fc6d65bfe6e4ed4f8fb2b5 Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.923366 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:27 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:27 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:27 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.923631 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:27 crc kubenswrapper[4931]: I1129 00:02:27.944171 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g72fb"] Nov 29 00:02:27 crc kubenswrapper[4931]: W1129 00:02:27.956462 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74999bec_b01b_4643_9cad_fa5e146089f5.slice/crio-21b11a34f36e3b7223945b2e07bc68750c563f5052dc415aba1f4a4b762f3d1f WatchSource:0}: Error finding container 21b11a34f36e3b7223945b2e07bc68750c563f5052dc415aba1f4a4b762f3d1f: Status 404 returned error can't find the container with id 21b11a34f36e3b7223945b2e07bc68750c563f5052dc415aba1f4a4b762f3d1f Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.022193 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2046696b-1636-47ba-9b22-d5eff643436c","Type":"ContainerStarted","Data":"13a319e8ea3e358ee4fb0475cbac8aa449efe3a9a1fc6d65bfe6e4ed4f8fb2b5"} Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.026400 4931 generic.go:334] "Generic (PLEG): container finished" podID="96bbcdc9-9b57-409e-b5b9-79a99034208a" containerID="a0faded80efe8587290a77ed570ea80cd569c557fbf016a8c3089016d33ab90b" exitCode=0 Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.026508 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" event={"ID":"96bbcdc9-9b57-409e-b5b9-79a99034208a","Type":"ContainerDied","Data":"a0faded80efe8587290a77ed570ea80cd569c557fbf016a8c3089016d33ab90b"} Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.029833 4931 generic.go:334] "Generic (PLEG): container finished" podID="2e2b7be6-d4e7-4e92-b419-bda83df88970" containerID="39b3341548aea9fbb9cc2e6501a0b834580047b5ec4d94faf1f3103146718472" exitCode=0 Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.030406 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9c9x" event={"ID":"2e2b7be6-d4e7-4e92-b419-bda83df88970","Type":"ContainerDied","Data":"39b3341548aea9fbb9cc2e6501a0b834580047b5ec4d94faf1f3103146718472"} Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.030432 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9c9x" event={"ID":"2e2b7be6-d4e7-4e92-b419-bda83df88970","Type":"ContainerStarted","Data":"cec7b0fa99c7095c785da0e358bc1d27ad2dfa3c30cd2ff81167ba8543e53ce3"} Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.059918 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g72fb" event={"ID":"74999bec-b01b-4643-9cad-fa5e146089f5","Type":"ContainerStarted","Data":"21b11a34f36e3b7223945b2e07bc68750c563f5052dc415aba1f4a4b762f3d1f"} Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.078044 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-rx5zl" Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.864794 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.868269 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.871204 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.871322 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.872875 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.924510 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:28 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:28 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:28 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.924586 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.974403 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30d33f4f-3912-4449-973f-5f66cbc00ad9-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"30d33f4f-3912-4449-973f-5f66cbc00ad9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 00:02:28 crc kubenswrapper[4931]: I1129 00:02:28.974514 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30d33f4f-3912-4449-973f-5f66cbc00ad9-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"30d33f4f-3912-4449-973f-5f66cbc00ad9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.073156 4931 generic.go:334] "Generic (PLEG): container finished" podID="74999bec-b01b-4643-9cad-fa5e146089f5" containerID="249487ee69bf04eccbc907c850cfcbba1b496887eee5a30df13961dc389d8fa3" exitCode=0 Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.073274 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g72fb" event={"ID":"74999bec-b01b-4643-9cad-fa5e146089f5","Type":"ContainerDied","Data":"249487ee69bf04eccbc907c850cfcbba1b496887eee5a30df13961dc389d8fa3"} Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.075033 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30d33f4f-3912-4449-973f-5f66cbc00ad9-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"30d33f4f-3912-4449-973f-5f66cbc00ad9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.075105 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30d33f4f-3912-4449-973f-5f66cbc00ad9-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"30d33f4f-3912-4449-973f-5f66cbc00ad9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.075191 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30d33f4f-3912-4449-973f-5f66cbc00ad9-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"30d33f4f-3912-4449-973f-5f66cbc00ad9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.076146 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2046696b-1636-47ba-9b22-d5eff643436c","Type":"ContainerStarted","Data":"5a4834280aeda4b2482bd27d0d643aafec78464e4b4437fe12ed01e7fbb4648b"} Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.098957 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30d33f4f-3912-4449-973f-5f66cbc00ad9-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"30d33f4f-3912-4449-973f-5f66cbc00ad9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.104171 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.10415548 podStartE2EDuration="3.10415548s" podCreationTimestamp="2025-11-29 00:02:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:02:29.102927964 +0000 UTC m=+162.264821216" watchObservedRunningTime="2025-11-29 00:02:29.10415548 +0000 UTC m=+162.266048712" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.190070 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.397861 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.401852 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.481277 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ck2r\" (UniqueName: \"kubernetes.io/projected/96bbcdc9-9b57-409e-b5b9-79a99034208a-kube-api-access-6ck2r\") pod \"96bbcdc9-9b57-409e-b5b9-79a99034208a\" (UID: \"96bbcdc9-9b57-409e-b5b9-79a99034208a\") " Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.481373 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96bbcdc9-9b57-409e-b5b9-79a99034208a-config-volume\") pod \"96bbcdc9-9b57-409e-b5b9-79a99034208a\" (UID: \"96bbcdc9-9b57-409e-b5b9-79a99034208a\") " Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.481398 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96bbcdc9-9b57-409e-b5b9-79a99034208a-secret-volume\") pod \"96bbcdc9-9b57-409e-b5b9-79a99034208a\" (UID: \"96bbcdc9-9b57-409e-b5b9-79a99034208a\") " Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.482153 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96bbcdc9-9b57-409e-b5b9-79a99034208a-config-volume" (OuterVolumeSpecName: "config-volume") pod "96bbcdc9-9b57-409e-b5b9-79a99034208a" (UID: "96bbcdc9-9b57-409e-b5b9-79a99034208a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.485989 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96bbcdc9-9b57-409e-b5b9-79a99034208a-kube-api-access-6ck2r" (OuterVolumeSpecName: "kube-api-access-6ck2r") pod "96bbcdc9-9b57-409e-b5b9-79a99034208a" (UID: "96bbcdc9-9b57-409e-b5b9-79a99034208a"). InnerVolumeSpecName "kube-api-access-6ck2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.486859 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96bbcdc9-9b57-409e-b5b9-79a99034208a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "96bbcdc9-9b57-409e-b5b9-79a99034208a" (UID: "96bbcdc9-9b57-409e-b5b9-79a99034208a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.582397 4931 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96bbcdc9-9b57-409e-b5b9-79a99034208a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.582424 4931 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96bbcdc9-9b57-409e-b5b9-79a99034208a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.582435 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ck2r\" (UniqueName: \"kubernetes.io/projected/96bbcdc9-9b57-409e-b5b9-79a99034208a-kube-api-access-6ck2r\") on node \"crc\" DevicePath \"\"" Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.924795 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:29 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:29 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:29 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:29 crc kubenswrapper[4931]: I1129 00:02:29.924869 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:30 crc kubenswrapper[4931]: I1129 00:02:30.089885 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" Nov 29 00:02:30 crc kubenswrapper[4931]: I1129 00:02:30.090612 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7" event={"ID":"96bbcdc9-9b57-409e-b5b9-79a99034208a","Type":"ContainerDied","Data":"4da6124bf83dca21ab93286bffc7edaf51f2333742a5d124c8169da9cd027cb0"} Nov 29 00:02:30 crc kubenswrapper[4931]: I1129 00:02:30.090652 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4da6124bf83dca21ab93286bffc7edaf51f2333742a5d124c8169da9cd027cb0" Nov 29 00:02:30 crc kubenswrapper[4931]: I1129 00:02:30.095309 4931 generic.go:334] "Generic (PLEG): container finished" podID="2046696b-1636-47ba-9b22-d5eff643436c" containerID="5a4834280aeda4b2482bd27d0d643aafec78464e4b4437fe12ed01e7fbb4648b" exitCode=0 Nov 29 00:02:30 crc kubenswrapper[4931]: I1129 00:02:30.095445 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2046696b-1636-47ba-9b22-d5eff643436c","Type":"ContainerDied","Data":"5a4834280aeda4b2482bd27d0d643aafec78464e4b4437fe12ed01e7fbb4648b"} Nov 29 00:02:30 crc kubenswrapper[4931]: I1129 00:02:30.096996 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"30d33f4f-3912-4449-973f-5f66cbc00ad9","Type":"ContainerStarted","Data":"3ae2698a158505544e44442d5d09e0bbc56c170359530db4bc63704132ce5b59"} Nov 29 00:02:30 crc kubenswrapper[4931]: I1129 00:02:30.923667 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:30 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:30 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:30 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:30 crc kubenswrapper[4931]: I1129 00:02:30.924003 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:31 crc kubenswrapper[4931]: I1129 00:02:31.116246 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"30d33f4f-3912-4449-973f-5f66cbc00ad9","Type":"ContainerStarted","Data":"9e414b14c5ad87cffac5823f13708df0c3d52e48bf99894882308972eb8976b4"} Nov 29 00:02:31 crc kubenswrapper[4931]: I1129 00:02:31.400405 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 00:02:31 crc kubenswrapper[4931]: I1129 00:02:31.413846 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2046696b-1636-47ba-9b22-d5eff643436c-kube-api-access\") pod \"2046696b-1636-47ba-9b22-d5eff643436c\" (UID: \"2046696b-1636-47ba-9b22-d5eff643436c\") " Nov 29 00:02:31 crc kubenswrapper[4931]: I1129 00:02:31.414036 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2046696b-1636-47ba-9b22-d5eff643436c-kubelet-dir\") pod \"2046696b-1636-47ba-9b22-d5eff643436c\" (UID: \"2046696b-1636-47ba-9b22-d5eff643436c\") " Nov 29 00:02:31 crc kubenswrapper[4931]: I1129 00:02:31.414376 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2046696b-1636-47ba-9b22-d5eff643436c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2046696b-1636-47ba-9b22-d5eff643436c" (UID: "2046696b-1636-47ba-9b22-d5eff643436c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:02:31 crc kubenswrapper[4931]: I1129 00:02:31.424037 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2046696b-1636-47ba-9b22-d5eff643436c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2046696b-1636-47ba-9b22-d5eff643436c" (UID: "2046696b-1636-47ba-9b22-d5eff643436c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:02:31 crc kubenswrapper[4931]: I1129 00:02:31.516591 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2046696b-1636-47ba-9b22-d5eff643436c-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 00:02:31 crc kubenswrapper[4931]: I1129 00:02:31.516644 4931 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2046696b-1636-47ba-9b22-d5eff643436c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 29 00:02:31 crc kubenswrapper[4931]: I1129 00:02:31.924406 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:31 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:31 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:31 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:31 crc kubenswrapper[4931]: I1129 00:02:31.924521 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:32 crc kubenswrapper[4931]: I1129 00:02:32.125938 4931 generic.go:334] "Generic (PLEG): container finished" podID="30d33f4f-3912-4449-973f-5f66cbc00ad9" containerID="9e414b14c5ad87cffac5823f13708df0c3d52e48bf99894882308972eb8976b4" exitCode=0 Nov 29 00:02:32 crc kubenswrapper[4931]: I1129 00:02:32.126005 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"30d33f4f-3912-4449-973f-5f66cbc00ad9","Type":"ContainerDied","Data":"9e414b14c5ad87cffac5823f13708df0c3d52e48bf99894882308972eb8976b4"} Nov 29 00:02:32 crc kubenswrapper[4931]: I1129 00:02:32.129226 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2046696b-1636-47ba-9b22-d5eff643436c","Type":"ContainerDied","Data":"13a319e8ea3e358ee4fb0475cbac8aa449efe3a9a1fc6d65bfe6e4ed4f8fb2b5"} Nov 29 00:02:32 crc kubenswrapper[4931]: I1129 00:02:32.129250 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13a319e8ea3e358ee4fb0475cbac8aa449efe3a9a1fc6d65bfe6e4ed4f8fb2b5" Nov 29 00:02:32 crc kubenswrapper[4931]: I1129 00:02:32.129304 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 29 00:02:32 crc kubenswrapper[4931]: I1129 00:02:32.921259 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-hbmrd" Nov 29 00:02:32 crc kubenswrapper[4931]: I1129 00:02:32.927663 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:32 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:32 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:32 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:32 crc kubenswrapper[4931]: I1129 00:02:32.927871 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:33 crc kubenswrapper[4931]: I1129 00:02:33.923265 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:33 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:33 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:33 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:33 crc kubenswrapper[4931]: I1129 00:02:33.923577 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:34 crc kubenswrapper[4931]: I1129 00:02:34.924046 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:34 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:34 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:34 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:34 crc kubenswrapper[4931]: I1129 00:02:34.924134 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:35 crc kubenswrapper[4931]: I1129 00:02:35.922831 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:35 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:35 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:35 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:35 crc kubenswrapper[4931]: I1129 00:02:35.922920 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:36 crc kubenswrapper[4931]: I1129 00:02:36.796901 4931 patch_prober.go:28] interesting pod/console-f9d7485db-vvmnt container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 29 00:02:36 crc kubenswrapper[4931]: I1129 00:02:36.797657 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vvmnt" podUID="2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 29 00:02:36 crc kubenswrapper[4931]: I1129 00:02:36.924493 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:36 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:36 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:36 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:36 crc kubenswrapper[4931]: I1129 00:02:36.924581 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:37 crc kubenswrapper[4931]: I1129 00:02:37.258688 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:02:37 crc kubenswrapper[4931]: I1129 00:02:37.265288 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e38ad69-f5fb-415d-ad76-e899aed032a6-metrics-certs\") pod \"network-metrics-daemon-7dwtb\" (UID: \"5e38ad69-f5fb-415d-ad76-e899aed032a6\") " pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:02:37 crc kubenswrapper[4931]: I1129 00:02:37.458618 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7dwtb" Nov 29 00:02:37 crc kubenswrapper[4931]: I1129 00:02:37.924249 4931 patch_prober.go:28] interesting pod/router-default-5444994796-plmcw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 29 00:02:37 crc kubenswrapper[4931]: [-]has-synced failed: reason withheld Nov 29 00:02:37 crc kubenswrapper[4931]: [+]process-running ok Nov 29 00:02:37 crc kubenswrapper[4931]: healthz check failed Nov 29 00:02:37 crc kubenswrapper[4931]: I1129 00:02:37.924470 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-plmcw" podUID="bd34298a-7f8c-4264-97d5-4503c5182e49" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 29 00:02:38 crc kubenswrapper[4931]: I1129 00:02:38.686171 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 00:02:38 crc kubenswrapper[4931]: I1129 00:02:38.782415 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30d33f4f-3912-4449-973f-5f66cbc00ad9-kube-api-access\") pod \"30d33f4f-3912-4449-973f-5f66cbc00ad9\" (UID: \"30d33f4f-3912-4449-973f-5f66cbc00ad9\") " Nov 29 00:02:38 crc kubenswrapper[4931]: I1129 00:02:38.782464 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30d33f4f-3912-4449-973f-5f66cbc00ad9-kubelet-dir\") pod \"30d33f4f-3912-4449-973f-5f66cbc00ad9\" (UID: \"30d33f4f-3912-4449-973f-5f66cbc00ad9\") " Nov 29 00:02:38 crc kubenswrapper[4931]: I1129 00:02:38.782843 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/30d33f4f-3912-4449-973f-5f66cbc00ad9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "30d33f4f-3912-4449-973f-5f66cbc00ad9" (UID: "30d33f4f-3912-4449-973f-5f66cbc00ad9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:02:38 crc kubenswrapper[4931]: I1129 00:02:38.792013 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30d33f4f-3912-4449-973f-5f66cbc00ad9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "30d33f4f-3912-4449-973f-5f66cbc00ad9" (UID: "30d33f4f-3912-4449-973f-5f66cbc00ad9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:02:38 crc kubenswrapper[4931]: I1129 00:02:38.884451 4931 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30d33f4f-3912-4449-973f-5f66cbc00ad9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 29 00:02:38 crc kubenswrapper[4931]: I1129 00:02:38.884494 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30d33f4f-3912-4449-973f-5f66cbc00ad9-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 00:02:38 crc kubenswrapper[4931]: I1129 00:02:38.924084 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:38 crc kubenswrapper[4931]: I1129 00:02:38.927149 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-plmcw" Nov 29 00:02:39 crc kubenswrapper[4931]: I1129 00:02:39.207378 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"30d33f4f-3912-4449-973f-5f66cbc00ad9","Type":"ContainerDied","Data":"3ae2698a158505544e44442d5d09e0bbc56c170359530db4bc63704132ce5b59"} Nov 29 00:02:39 crc kubenswrapper[4931]: I1129 00:02:39.207704 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ae2698a158505544e44442d5d09e0bbc56c170359530db4bc63704132ce5b59" Nov 29 00:02:39 crc kubenswrapper[4931]: I1129 00:02:39.207431 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 29 00:02:43 crc kubenswrapper[4931]: I1129 00:02:43.780137 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:02:46 crc kubenswrapper[4931]: I1129 00:02:46.803263 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:46 crc kubenswrapper[4931]: I1129 00:02:46.809712 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:02:52 crc kubenswrapper[4931]: I1129 00:02:52.284290 4931 generic.go:334] "Generic (PLEG): container finished" podID="35311d49-bd1d-4351-90a8-f332a6c3a5d2" containerID="b1627da317dd8fc82d7909fa64c1b799c974d40852b9e2c0e88f504805d294f4" exitCode=0 Nov 29 00:02:52 crc kubenswrapper[4931]: I1129 00:02:52.284388 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29406240-4vs86" event={"ID":"35311d49-bd1d-4351-90a8-f332a6c3a5d2","Type":"ContainerDied","Data":"b1627da317dd8fc82d7909fa64c1b799c974d40852b9e2c0e88f504805d294f4"} Nov 29 00:02:53 crc kubenswrapper[4931]: I1129 00:02:53.735647 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:02:53 crc kubenswrapper[4931]: I1129 00:02:53.735955 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:02:55 crc kubenswrapper[4931]: I1129 00:02:55.243637 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 29 00:02:57 crc kubenswrapper[4931]: I1129 00:02:57.274226 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xdssh" Nov 29 00:02:57 crc kubenswrapper[4931]: E1129 00:02:57.615425 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 29 00:02:57 crc kubenswrapper[4931]: E1129 00:02:57.615640 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2f65v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-r5h8c_openshift-marketplace(39557253-7dac-4846-a945-44b03824ec08): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 00:02:57 crc kubenswrapper[4931]: E1129 00:02:57.616773 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-r5h8c" podUID="39557253-7dac-4846-a945-44b03824ec08" Nov 29 00:02:58 crc kubenswrapper[4931]: E1129 00:02:58.016454 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 29 00:02:58 crc kubenswrapper[4931]: E1129 00:02:58.016616 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6fq6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-hzhd2_openshift-marketplace(7c5af5e4-0d88-443e-8f17-93ce08cb4bab): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 00:02:58 crc kubenswrapper[4931]: E1129 00:02:58.017895 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-hzhd2" podUID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" Nov 29 00:03:03 crc kubenswrapper[4931]: E1129 00:03:03.043226 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-hzhd2" podUID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" Nov 29 00:03:03 crc kubenswrapper[4931]: E1129 00:03:03.045011 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-r5h8c" podUID="39557253-7dac-4846-a945-44b03824ec08" Nov 29 00:03:03 crc kubenswrapper[4931]: E1129 00:03:03.121966 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 29 00:03:03 crc kubenswrapper[4931]: E1129 00:03:03.122210 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d92sp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-cdzck_openshift-marketplace(3e56db6c-4dd6-461b-8f37-5ce71ea285b3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 00:03:03 crc kubenswrapper[4931]: E1129 00:03:03.123434 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-cdzck" podUID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" Nov 29 00:03:03 crc kubenswrapper[4931]: E1129 00:03:03.135029 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 29 00:03:03 crc kubenswrapper[4931]: E1129 00:03:03.135150 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-26p65,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6ngbj_openshift-marketplace(bf2c190f-062c-407a-87dd-48a5c4ac07d5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 00:03:03 crc kubenswrapper[4931]: E1129 00:03:03.136578 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6ngbj" podUID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" Nov 29 00:03:03 crc kubenswrapper[4931]: E1129 00:03:03.172328 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 29 00:03:03 crc kubenswrapper[4931]: E1129 00:03:03.172643 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pqf6h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bk72v_openshift-marketplace(97c7b21f-8584-482c-b18b-074549e1a120): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 00:03:03 crc kubenswrapper[4931]: E1129 00:03:03.174056 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bk72v" podUID="97c7b21f-8584-482c-b18b-074549e1a120" Nov 29 00:03:05 crc kubenswrapper[4931]: E1129 00:03:05.908485 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bk72v" podUID="97c7b21f-8584-482c-b18b-074549e1a120" Nov 29 00:03:05 crc kubenswrapper[4931]: E1129 00:03:05.908672 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6ngbj" podUID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" Nov 29 00:03:05 crc kubenswrapper[4931]: E1129 00:03:05.908748 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-cdzck" podUID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" Nov 29 00:03:05 crc kubenswrapper[4931]: E1129 00:03:05.941669 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 29 00:03:05 crc kubenswrapper[4931]: E1129 00:03:05.941851 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p6f2s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-q9c9x_openshift-marketplace(2e2b7be6-d4e7-4e92-b419-bda83df88970): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 00:03:05 crc kubenswrapper[4931]: E1129 00:03:05.943456 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-q9c9x" podUID="2e2b7be6-d4e7-4e92-b419-bda83df88970" Nov 29 00:03:05 crc kubenswrapper[4931]: E1129 00:03:05.951892 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 29 00:03:05 crc kubenswrapper[4931]: E1129 00:03:05.952054 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5vmjf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-g72fb_openshift-marketplace(74999bec-b01b-4643-9cad-fa5e146089f5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 29 00:03:05 crc kubenswrapper[4931]: E1129 00:03:05.953334 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-g72fb" podUID="74999bec-b01b-4643-9cad-fa5e146089f5" Nov 29 00:03:05 crc kubenswrapper[4931]: I1129 00:03:05.964121 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29406240-4vs86" Nov 29 00:03:06 crc kubenswrapper[4931]: I1129 00:03:06.096913 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/35311d49-bd1d-4351-90a8-f332a6c3a5d2-serviceca\") pod \"35311d49-bd1d-4351-90a8-f332a6c3a5d2\" (UID: \"35311d49-bd1d-4351-90a8-f332a6c3a5d2\") " Nov 29 00:03:06 crc kubenswrapper[4931]: I1129 00:03:06.097249 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft4tc\" (UniqueName: \"kubernetes.io/projected/35311d49-bd1d-4351-90a8-f332a6c3a5d2-kube-api-access-ft4tc\") pod \"35311d49-bd1d-4351-90a8-f332a6c3a5d2\" (UID: \"35311d49-bd1d-4351-90a8-f332a6c3a5d2\") " Nov 29 00:03:06 crc kubenswrapper[4931]: I1129 00:03:06.098712 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35311d49-bd1d-4351-90a8-f332a6c3a5d2-serviceca" (OuterVolumeSpecName: "serviceca") pod "35311d49-bd1d-4351-90a8-f332a6c3a5d2" (UID: "35311d49-bd1d-4351-90a8-f332a6c3a5d2"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:03:06 crc kubenswrapper[4931]: I1129 00:03:06.106200 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35311d49-bd1d-4351-90a8-f332a6c3a5d2-kube-api-access-ft4tc" (OuterVolumeSpecName: "kube-api-access-ft4tc") pod "35311d49-bd1d-4351-90a8-f332a6c3a5d2" (UID: "35311d49-bd1d-4351-90a8-f332a6c3a5d2"). InnerVolumeSpecName "kube-api-access-ft4tc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:03:06 crc kubenswrapper[4931]: I1129 00:03:06.199325 4931 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/35311d49-bd1d-4351-90a8-f332a6c3a5d2-serviceca\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:06 crc kubenswrapper[4931]: I1129 00:03:06.199372 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft4tc\" (UniqueName: \"kubernetes.io/projected/35311d49-bd1d-4351-90a8-f332a6c3a5d2-kube-api-access-ft4tc\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:06 crc kubenswrapper[4931]: I1129 00:03:06.398424 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2tsz" event={"ID":"2f5b8171-a997-49fe-ac00-eb90258f747a","Type":"ContainerStarted","Data":"8d66ea5f5427eb66587c076b58fc47a9034ef5091990e6fc5f68d46257df6d06"} Nov 29 00:03:06 crc kubenswrapper[4931]: I1129 00:03:06.400140 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7dwtb"] Nov 29 00:03:06 crc kubenswrapper[4931]: I1129 00:03:06.403763 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29406240-4vs86" event={"ID":"35311d49-bd1d-4351-90a8-f332a6c3a5d2","Type":"ContainerDied","Data":"641c61212b90b06b20e26399aaf0090e858e8f40bd3eb7ec570937f515cb5926"} Nov 29 00:03:06 crc kubenswrapper[4931]: I1129 00:03:06.403798 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="641c61212b90b06b20e26399aaf0090e858e8f40bd3eb7ec570937f515cb5926" Nov 29 00:03:06 crc kubenswrapper[4931]: I1129 00:03:06.403868 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29406240-4vs86" Nov 29 00:03:06 crc kubenswrapper[4931]: E1129 00:03:06.407862 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-g72fb" podUID="74999bec-b01b-4643-9cad-fa5e146089f5" Nov 29 00:03:06 crc kubenswrapper[4931]: E1129 00:03:06.408097 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-q9c9x" podUID="2e2b7be6-d4e7-4e92-b419-bda83df88970" Nov 29 00:03:07 crc kubenswrapper[4931]: I1129 00:03:07.415088 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" event={"ID":"5e38ad69-f5fb-415d-ad76-e899aed032a6","Type":"ContainerStarted","Data":"6c2836f3a177d03b318e82a7588a630f44d32c4eda3d45b998440425ea81b566"} Nov 29 00:03:07 crc kubenswrapper[4931]: I1129 00:03:07.417356 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" event={"ID":"5e38ad69-f5fb-415d-ad76-e899aed032a6","Type":"ContainerStarted","Data":"09bb742d83b442d1fddc9537f90085616f0e1ac4a3feee0d2fcc2e3005dc872d"} Nov 29 00:03:07 crc kubenswrapper[4931]: I1129 00:03:07.417581 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7dwtb" event={"ID":"5e38ad69-f5fb-415d-ad76-e899aed032a6","Type":"ContainerStarted","Data":"60d10eb0f04e7cef3caf6d49d6f5ea0c7a70d847874536b4b8bc7b5cb29673f8"} Nov 29 00:03:07 crc kubenswrapper[4931]: I1129 00:03:07.433827 4931 generic.go:334] "Generic (PLEG): container finished" podID="2f5b8171-a997-49fe-ac00-eb90258f747a" containerID="8d66ea5f5427eb66587c076b58fc47a9034ef5091990e6fc5f68d46257df6d06" exitCode=0 Nov 29 00:03:07 crc kubenswrapper[4931]: I1129 00:03:07.433867 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2tsz" event={"ID":"2f5b8171-a997-49fe-ac00-eb90258f747a","Type":"ContainerDied","Data":"8d66ea5f5427eb66587c076b58fc47a9034ef5091990e6fc5f68d46257df6d06"} Nov 29 00:03:07 crc kubenswrapper[4931]: I1129 00:03:07.433897 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2tsz" event={"ID":"2f5b8171-a997-49fe-ac00-eb90258f747a","Type":"ContainerStarted","Data":"69999e8efc16d246c63b1b1717e57fab2a6d6267ed8dfae22ce6e85640697570"} Nov 29 00:03:07 crc kubenswrapper[4931]: I1129 00:03:07.454418 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-7dwtb" podStartSLOduration=174.454383045 podStartE2EDuration="2m54.454383045s" podCreationTimestamp="2025-11-29 00:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:03:07.442716284 +0000 UTC m=+200.604609536" watchObservedRunningTime="2025-11-29 00:03:07.454383045 +0000 UTC m=+200.616276307" Nov 29 00:03:07 crc kubenswrapper[4931]: I1129 00:03:07.483560 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p2tsz" podStartSLOduration=2.4517559589999998 podStartE2EDuration="42.483537667s" podCreationTimestamp="2025-11-29 00:02:25 +0000 UTC" firstStartedPulling="2025-11-29 00:02:27.006865953 +0000 UTC m=+160.168759185" lastFinishedPulling="2025-11-29 00:03:07.038647651 +0000 UTC m=+200.200540893" observedRunningTime="2025-11-29 00:03:07.47886616 +0000 UTC m=+200.640759422" watchObservedRunningTime="2025-11-29 00:03:07.483537667 +0000 UTC m=+200.645430939" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.564071 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 29 00:03:08 crc kubenswrapper[4931]: E1129 00:03:08.564998 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35311d49-bd1d-4351-90a8-f332a6c3a5d2" containerName="image-pruner" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.565020 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="35311d49-bd1d-4351-90a8-f332a6c3a5d2" containerName="image-pruner" Nov 29 00:03:08 crc kubenswrapper[4931]: E1129 00:03:08.565042 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2046696b-1636-47ba-9b22-d5eff643436c" containerName="pruner" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.565055 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2046696b-1636-47ba-9b22-d5eff643436c" containerName="pruner" Nov 29 00:03:08 crc kubenswrapper[4931]: E1129 00:03:08.565071 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96bbcdc9-9b57-409e-b5b9-79a99034208a" containerName="collect-profiles" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.565082 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="96bbcdc9-9b57-409e-b5b9-79a99034208a" containerName="collect-profiles" Nov 29 00:03:08 crc kubenswrapper[4931]: E1129 00:03:08.565113 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d33f4f-3912-4449-973f-5f66cbc00ad9" containerName="pruner" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.565124 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d33f4f-3912-4449-973f-5f66cbc00ad9" containerName="pruner" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.565284 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="30d33f4f-3912-4449-973f-5f66cbc00ad9" containerName="pruner" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.565308 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2046696b-1636-47ba-9b22-d5eff643436c" containerName="pruner" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.565320 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="96bbcdc9-9b57-409e-b5b9-79a99034208a" containerName="collect-profiles" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.565337 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="35311d49-bd1d-4351-90a8-f332a6c3a5d2" containerName="image-pruner" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.565988 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.569988 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.570049 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.582613 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.738925 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e67aff2-bf6c-4f67-922a-c0454e7026ff-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9e67aff2-bf6c-4f67-922a-c0454e7026ff\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.739062 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e67aff2-bf6c-4f67-922a-c0454e7026ff-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9e67aff2-bf6c-4f67-922a-c0454e7026ff\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.840457 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e67aff2-bf6c-4f67-922a-c0454e7026ff-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9e67aff2-bf6c-4f67-922a-c0454e7026ff\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.840547 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e67aff2-bf6c-4f67-922a-c0454e7026ff-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9e67aff2-bf6c-4f67-922a-c0454e7026ff\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.841149 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e67aff2-bf6c-4f67-922a-c0454e7026ff-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9e67aff2-bf6c-4f67-922a-c0454e7026ff\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.861560 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e67aff2-bf6c-4f67-922a-c0454e7026ff-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9e67aff2-bf6c-4f67-922a-c0454e7026ff\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 00:03:08 crc kubenswrapper[4931]: I1129 00:03:08.891900 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 00:03:09 crc kubenswrapper[4931]: I1129 00:03:09.349947 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 29 00:03:09 crc kubenswrapper[4931]: W1129 00:03:09.358998 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9e67aff2_bf6c_4f67_922a_c0454e7026ff.slice/crio-887522ea113eb67e6a261e9761efaaaea8e488a31a4710c7f6f0f02192512c01 WatchSource:0}: Error finding container 887522ea113eb67e6a261e9761efaaaea8e488a31a4710c7f6f0f02192512c01: Status 404 returned error can't find the container with id 887522ea113eb67e6a261e9761efaaaea8e488a31a4710c7f6f0f02192512c01 Nov 29 00:03:09 crc kubenswrapper[4931]: I1129 00:03:09.444858 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9e67aff2-bf6c-4f67-922a-c0454e7026ff","Type":"ContainerStarted","Data":"887522ea113eb67e6a261e9761efaaaea8e488a31a4710c7f6f0f02192512c01"} Nov 29 00:03:11 crc kubenswrapper[4931]: I1129 00:03:11.136276 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9e67aff2-bf6c-4f67-922a-c0454e7026ff","Type":"ContainerStarted","Data":"1dcb23af2ca68a7eaf20abe4927241cc20c994431f4d7a036fe742b0d2aba2f7"} Nov 29 00:03:11 crc kubenswrapper[4931]: I1129 00:03:11.151560 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=3.151534871 podStartE2EDuration="3.151534871s" podCreationTimestamp="2025-11-29 00:03:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:03:11.150782809 +0000 UTC m=+204.312676051" watchObservedRunningTime="2025-11-29 00:03:11.151534871 +0000 UTC m=+204.313428143" Nov 29 00:03:12 crc kubenswrapper[4931]: I1129 00:03:12.144151 4931 generic.go:334] "Generic (PLEG): container finished" podID="9e67aff2-bf6c-4f67-922a-c0454e7026ff" containerID="1dcb23af2ca68a7eaf20abe4927241cc20c994431f4d7a036fe742b0d2aba2f7" exitCode=0 Nov 29 00:03:12 crc kubenswrapper[4931]: I1129 00:03:12.144200 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9e67aff2-bf6c-4f67-922a-c0454e7026ff","Type":"ContainerDied","Data":"1dcb23af2ca68a7eaf20abe4927241cc20c994431f4d7a036fe742b0d2aba2f7"} Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.369092 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.370472 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.379291 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.540281 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-var-lock\") pod \"installer-9-crc\" (UID: \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.540396 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-kube-api-access\") pod \"installer-9-crc\" (UID: \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.540457 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-kubelet-dir\") pod \"installer-9-crc\" (UID: \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.573042 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.642171 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e67aff2-bf6c-4f67-922a-c0454e7026ff-kubelet-dir\") pod \"9e67aff2-bf6c-4f67-922a-c0454e7026ff\" (UID: \"9e67aff2-bf6c-4f67-922a-c0454e7026ff\") " Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.642217 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e67aff2-bf6c-4f67-922a-c0454e7026ff-kube-api-access\") pod \"9e67aff2-bf6c-4f67-922a-c0454e7026ff\" (UID: \"9e67aff2-bf6c-4f67-922a-c0454e7026ff\") " Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.642349 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-kube-api-access\") pod \"installer-9-crc\" (UID: \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.642386 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-kubelet-dir\") pod \"installer-9-crc\" (UID: \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.642410 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-var-lock\") pod \"installer-9-crc\" (UID: \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.642499 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-var-lock\") pod \"installer-9-crc\" (UID: \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.642553 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e67aff2-bf6c-4f67-922a-c0454e7026ff-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9e67aff2-bf6c-4f67-922a-c0454e7026ff" (UID: "9e67aff2-bf6c-4f67-922a-c0454e7026ff"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.643420 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-kubelet-dir\") pod \"installer-9-crc\" (UID: \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.661055 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e67aff2-bf6c-4f67-922a-c0454e7026ff-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9e67aff2-bf6c-4f67-922a-c0454e7026ff" (UID: "9e67aff2-bf6c-4f67-922a-c0454e7026ff"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.690570 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-kube-api-access\") pod \"installer-9-crc\" (UID: \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.706324 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.743235 4931 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e67aff2-bf6c-4f67-922a-c0454e7026ff-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:13 crc kubenswrapper[4931]: I1129 00:03:13.743263 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e67aff2-bf6c-4f67-922a-c0454e7026ff-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:14 crc kubenswrapper[4931]: I1129 00:03:14.097465 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 29 00:03:14 crc kubenswrapper[4931]: I1129 00:03:14.160162 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"4b0e513e-266b-4aeb-9c04-e7b07d28ec19","Type":"ContainerStarted","Data":"44f0898bd30a954dd7b54375f20822ecafc4eb76d8106d4f9cc7fce3682eddde"} Nov 29 00:03:14 crc kubenswrapper[4931]: I1129 00:03:14.162297 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9e67aff2-bf6c-4f67-922a-c0454e7026ff","Type":"ContainerDied","Data":"887522ea113eb67e6a261e9761efaaaea8e488a31a4710c7f6f0f02192512c01"} Nov 29 00:03:14 crc kubenswrapper[4931]: I1129 00:03:14.162320 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="887522ea113eb67e6a261e9761efaaaea8e488a31a4710c7f6f0f02192512c01" Nov 29 00:03:14 crc kubenswrapper[4931]: I1129 00:03:14.162407 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 29 00:03:16 crc kubenswrapper[4931]: I1129 00:03:16.180493 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5h8c" event={"ID":"39557253-7dac-4846-a945-44b03824ec08","Type":"ContainerStarted","Data":"37f9cb08e81eb8bf7642bfd1d7d203d9ca104b67b40893273731704e39148fdd"} Nov 29 00:03:16 crc kubenswrapper[4931]: I1129 00:03:16.182769 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"4b0e513e-266b-4aeb-9c04-e7b07d28ec19","Type":"ContainerStarted","Data":"b36f834a1c60b2b466fb494d417c6565be12b8ac9d07e89bd8693fea59009422"} Nov 29 00:03:16 crc kubenswrapper[4931]: I1129 00:03:16.214276 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.214259467 podStartE2EDuration="3.214259467s" podCreationTimestamp="2025-11-29 00:03:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:03:16.210894629 +0000 UTC m=+209.372787871" watchObservedRunningTime="2025-11-29 00:03:16.214259467 +0000 UTC m=+209.376152699" Nov 29 00:03:16 crc kubenswrapper[4931]: I1129 00:03:16.239061 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:03:16 crc kubenswrapper[4931]: I1129 00:03:16.239116 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:03:16 crc kubenswrapper[4931]: I1129 00:03:16.316779 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:03:17 crc kubenswrapper[4931]: I1129 00:03:17.191275 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ngbj" event={"ID":"bf2c190f-062c-407a-87dd-48a5c4ac07d5","Type":"ContainerStarted","Data":"9e85638a6b3e08156edc042510cb95e75810f327c1d8f01d51313e3f8e9a1893"} Nov 29 00:03:17 crc kubenswrapper[4931]: I1129 00:03:17.193277 4931 generic.go:334] "Generic (PLEG): container finished" podID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" containerID="d996a4e5ae36ffe8bad8ef0312f727e651e25d79e1c6e5cb0a652084b08c3a65" exitCode=0 Nov 29 00:03:17 crc kubenswrapper[4931]: I1129 00:03:17.193329 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzhd2" event={"ID":"7c5af5e4-0d88-443e-8f17-93ce08cb4bab","Type":"ContainerDied","Data":"d996a4e5ae36ffe8bad8ef0312f727e651e25d79e1c6e5cb0a652084b08c3a65"} Nov 29 00:03:17 crc kubenswrapper[4931]: I1129 00:03:17.198305 4931 generic.go:334] "Generic (PLEG): container finished" podID="39557253-7dac-4846-a945-44b03824ec08" containerID="37f9cb08e81eb8bf7642bfd1d7d203d9ca104b67b40893273731704e39148fdd" exitCode=0 Nov 29 00:03:17 crc kubenswrapper[4931]: I1129 00:03:17.198349 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5h8c" event={"ID":"39557253-7dac-4846-a945-44b03824ec08","Type":"ContainerDied","Data":"37f9cb08e81eb8bf7642bfd1d7d203d9ca104b67b40893273731704e39148fdd"} Nov 29 00:03:17 crc kubenswrapper[4931]: I1129 00:03:17.241569 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:03:18 crc kubenswrapper[4931]: I1129 00:03:18.205522 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5h8c" event={"ID":"39557253-7dac-4846-a945-44b03824ec08","Type":"ContainerStarted","Data":"3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573"} Nov 29 00:03:18 crc kubenswrapper[4931]: I1129 00:03:18.208111 4931 generic.go:334] "Generic (PLEG): container finished" podID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" containerID="9e85638a6b3e08156edc042510cb95e75810f327c1d8f01d51313e3f8e9a1893" exitCode=0 Nov 29 00:03:18 crc kubenswrapper[4931]: I1129 00:03:18.208173 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ngbj" event={"ID":"bf2c190f-062c-407a-87dd-48a5c4ac07d5","Type":"ContainerDied","Data":"9e85638a6b3e08156edc042510cb95e75810f327c1d8f01d51313e3f8e9a1893"} Nov 29 00:03:18 crc kubenswrapper[4931]: I1129 00:03:18.212210 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzhd2" event={"ID":"7c5af5e4-0d88-443e-8f17-93ce08cb4bab","Type":"ContainerStarted","Data":"51bd0d1536062b0b9e4985ed7f272d8a71ee9ad94693d67ef731463be9e3532a"} Nov 29 00:03:18 crc kubenswrapper[4931]: I1129 00:03:18.227350 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r5h8c" podStartSLOduration=2.429060057 podStartE2EDuration="55.227301869s" podCreationTimestamp="2025-11-29 00:02:23 +0000 UTC" firstStartedPulling="2025-11-29 00:02:24.94373186 +0000 UTC m=+158.105625092" lastFinishedPulling="2025-11-29 00:03:17.741973672 +0000 UTC m=+210.903866904" observedRunningTime="2025-11-29 00:03:18.224408795 +0000 UTC m=+211.386302047" watchObservedRunningTime="2025-11-29 00:03:18.227301869 +0000 UTC m=+211.389195101" Nov 29 00:03:18 crc kubenswrapper[4931]: I1129 00:03:18.244390 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hzhd2" podStartSLOduration=2.472101565 podStartE2EDuration="55.244370978s" podCreationTimestamp="2025-11-29 00:02:23 +0000 UTC" firstStartedPulling="2025-11-29 00:02:24.938064985 +0000 UTC m=+158.099958237" lastFinishedPulling="2025-11-29 00:03:17.710334418 +0000 UTC m=+210.872227650" observedRunningTime="2025-11-29 00:03:18.241372 +0000 UTC m=+211.403265242" watchObservedRunningTime="2025-11-29 00:03:18.244370978 +0000 UTC m=+211.406264210" Nov 29 00:03:20 crc kubenswrapper[4931]: I1129 00:03:20.223617 4931 generic.go:334] "Generic (PLEG): container finished" podID="97c7b21f-8584-482c-b18b-074549e1a120" containerID="a4a6bb2a3b6aad78d54d32e58766ff0a7e8de757329e3fd25bb455dd5f6b4d15" exitCode=0 Nov 29 00:03:20 crc kubenswrapper[4931]: I1129 00:03:20.223704 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk72v" event={"ID":"97c7b21f-8584-482c-b18b-074549e1a120","Type":"ContainerDied","Data":"a4a6bb2a3b6aad78d54d32e58766ff0a7e8de757329e3fd25bb455dd5f6b4d15"} Nov 29 00:03:20 crc kubenswrapper[4931]: I1129 00:03:20.228205 4931 generic.go:334] "Generic (PLEG): container finished" podID="74999bec-b01b-4643-9cad-fa5e146089f5" containerID="522829861b2063315b77d619f9c3c5482f97715d07bc8fbbf3c6a2014a5d708a" exitCode=0 Nov 29 00:03:20 crc kubenswrapper[4931]: I1129 00:03:20.228288 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g72fb" event={"ID":"74999bec-b01b-4643-9cad-fa5e146089f5","Type":"ContainerDied","Data":"522829861b2063315b77d619f9c3c5482f97715d07bc8fbbf3c6a2014a5d708a"} Nov 29 00:03:20 crc kubenswrapper[4931]: I1129 00:03:20.231432 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ngbj" event={"ID":"bf2c190f-062c-407a-87dd-48a5c4ac07d5","Type":"ContainerStarted","Data":"a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3"} Nov 29 00:03:20 crc kubenswrapper[4931]: I1129 00:03:20.234469 4931 generic.go:334] "Generic (PLEG): container finished" podID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" containerID="5ad8f53a6c80d568856ef481f45a9b2779cb337002ccf1987f8625e079519d54" exitCode=0 Nov 29 00:03:20 crc kubenswrapper[4931]: I1129 00:03:20.234508 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cdzck" event={"ID":"3e56db6c-4dd6-461b-8f37-5ce71ea285b3","Type":"ContainerDied","Data":"5ad8f53a6c80d568856ef481f45a9b2779cb337002ccf1987f8625e079519d54"} Nov 29 00:03:20 crc kubenswrapper[4931]: I1129 00:03:20.305296 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6ngbj" podStartSLOduration=2.698604844 podStartE2EDuration="56.305279019s" podCreationTimestamp="2025-11-29 00:02:24 +0000 UTC" firstStartedPulling="2025-11-29 00:02:25.986350398 +0000 UTC m=+159.148243630" lastFinishedPulling="2025-11-29 00:03:19.593024573 +0000 UTC m=+212.754917805" observedRunningTime="2025-11-29 00:03:20.301956492 +0000 UTC m=+213.463849754" watchObservedRunningTime="2025-11-29 00:03:20.305279019 +0000 UTC m=+213.467172241" Nov 29 00:03:21 crc kubenswrapper[4931]: I1129 00:03:21.242017 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk72v" event={"ID":"97c7b21f-8584-482c-b18b-074549e1a120","Type":"ContainerStarted","Data":"f84d5d1e015f1a40c54ba5b98bb105100740fe21f39c48211300dba5df23d988"} Nov 29 00:03:21 crc kubenswrapper[4931]: I1129 00:03:21.243832 4931 generic.go:334] "Generic (PLEG): container finished" podID="2e2b7be6-d4e7-4e92-b419-bda83df88970" containerID="88a7b08d2e0dc9d9c1ae3969d79d20fbdf95bfa0080d33085ce90d4db9a828f1" exitCode=0 Nov 29 00:03:21 crc kubenswrapper[4931]: I1129 00:03:21.243927 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9c9x" event={"ID":"2e2b7be6-d4e7-4e92-b419-bda83df88970","Type":"ContainerDied","Data":"88a7b08d2e0dc9d9c1ae3969d79d20fbdf95bfa0080d33085ce90d4db9a828f1"} Nov 29 00:03:21 crc kubenswrapper[4931]: I1129 00:03:21.246747 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cdzck" event={"ID":"3e56db6c-4dd6-461b-8f37-5ce71ea285b3","Type":"ContainerStarted","Data":"e94a1ee77ce918558bc0b088d0a891a5e2eec1ffabdf7c7396cec64f9b099925"} Nov 29 00:03:21 crc kubenswrapper[4931]: I1129 00:03:21.258356 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bk72v" podStartSLOduration=2.248198972 podStartE2EDuration="56.258339898s" podCreationTimestamp="2025-11-29 00:02:25 +0000 UTC" firstStartedPulling="2025-11-29 00:02:27.017890554 +0000 UTC m=+160.179783786" lastFinishedPulling="2025-11-29 00:03:21.02803148 +0000 UTC m=+214.189924712" observedRunningTime="2025-11-29 00:03:21.25772643 +0000 UTC m=+214.419619672" watchObservedRunningTime="2025-11-29 00:03:21.258339898 +0000 UTC m=+214.420233140" Nov 29 00:03:21 crc kubenswrapper[4931]: I1129 00:03:21.274586 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cdzck" podStartSLOduration=2.350638666 podStartE2EDuration="58.274571742s" podCreationTimestamp="2025-11-29 00:02:23 +0000 UTC" firstStartedPulling="2025-11-29 00:02:24.939107675 +0000 UTC m=+158.101000917" lastFinishedPulling="2025-11-29 00:03:20.863040751 +0000 UTC m=+214.024933993" observedRunningTime="2025-11-29 00:03:21.27143408 +0000 UTC m=+214.433327322" watchObservedRunningTime="2025-11-29 00:03:21.274571742 +0000 UTC m=+214.436464974" Nov 29 00:03:21 crc kubenswrapper[4931]: I1129 00:03:21.445299 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2tsz"] Nov 29 00:03:21 crc kubenswrapper[4931]: I1129 00:03:21.445889 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p2tsz" podUID="2f5b8171-a997-49fe-ac00-eb90258f747a" containerName="registry-server" containerID="cri-o://69999e8efc16d246c63b1b1717e57fab2a6d6267ed8dfae22ce6e85640697570" gracePeriod=2 Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.253995 4931 generic.go:334] "Generic (PLEG): container finished" podID="2f5b8171-a997-49fe-ac00-eb90258f747a" containerID="69999e8efc16d246c63b1b1717e57fab2a6d6267ed8dfae22ce6e85640697570" exitCode=0 Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.254077 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2tsz" event={"ID":"2f5b8171-a997-49fe-ac00-eb90258f747a","Type":"ContainerDied","Data":"69999e8efc16d246c63b1b1717e57fab2a6d6267ed8dfae22ce6e85640697570"} Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.256352 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9c9x" event={"ID":"2e2b7be6-d4e7-4e92-b419-bda83df88970","Type":"ContainerStarted","Data":"34168fbe1775d3925b41507a0d49d70ae8af2f7997262babf242e0c6d248bcd3"} Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.258599 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g72fb" event={"ID":"74999bec-b01b-4643-9cad-fa5e146089f5","Type":"ContainerStarted","Data":"b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74"} Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.276322 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q9c9x" podStartSLOduration=2.4940466580000002 podStartE2EDuration="56.276296393s" podCreationTimestamp="2025-11-29 00:02:26 +0000 UTC" firstStartedPulling="2025-11-29 00:02:28.032421954 +0000 UTC m=+161.194315186" lastFinishedPulling="2025-11-29 00:03:21.814671689 +0000 UTC m=+214.976564921" observedRunningTime="2025-11-29 00:03:22.273470861 +0000 UTC m=+215.435364093" watchObservedRunningTime="2025-11-29 00:03:22.276296393 +0000 UTC m=+215.438189655" Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.291117 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g72fb" podStartSLOduration=3.126761294 podStartE2EDuration="55.291097515s" podCreationTimestamp="2025-11-29 00:02:27 +0000 UTC" firstStartedPulling="2025-11-29 00:02:29.077186915 +0000 UTC m=+162.239080147" lastFinishedPulling="2025-11-29 00:03:21.241523136 +0000 UTC m=+214.403416368" observedRunningTime="2025-11-29 00:03:22.288111278 +0000 UTC m=+215.450004580" watchObservedRunningTime="2025-11-29 00:03:22.291097515 +0000 UTC m=+215.452990737" Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.788783 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.958703 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5b8171-a997-49fe-ac00-eb90258f747a-utilities\") pod \"2f5b8171-a997-49fe-ac00-eb90258f747a\" (UID: \"2f5b8171-a997-49fe-ac00-eb90258f747a\") " Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.958832 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5b8171-a997-49fe-ac00-eb90258f747a-catalog-content\") pod \"2f5b8171-a997-49fe-ac00-eb90258f747a\" (UID: \"2f5b8171-a997-49fe-ac00-eb90258f747a\") " Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.958875 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2s8w\" (UniqueName: \"kubernetes.io/projected/2f5b8171-a997-49fe-ac00-eb90258f747a-kube-api-access-n2s8w\") pod \"2f5b8171-a997-49fe-ac00-eb90258f747a\" (UID: \"2f5b8171-a997-49fe-ac00-eb90258f747a\") " Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.959514 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f5b8171-a997-49fe-ac00-eb90258f747a-utilities" (OuterVolumeSpecName: "utilities") pod "2f5b8171-a997-49fe-ac00-eb90258f747a" (UID: "2f5b8171-a997-49fe-ac00-eb90258f747a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.975577 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f5b8171-a997-49fe-ac00-eb90258f747a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f5b8171-a997-49fe-ac00-eb90258f747a" (UID: "2f5b8171-a997-49fe-ac00-eb90258f747a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:03:22 crc kubenswrapper[4931]: I1129 00:03:22.982230 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f5b8171-a997-49fe-ac00-eb90258f747a-kube-api-access-n2s8w" (OuterVolumeSpecName: "kube-api-access-n2s8w") pod "2f5b8171-a997-49fe-ac00-eb90258f747a" (UID: "2f5b8171-a997-49fe-ac00-eb90258f747a"). InnerVolumeSpecName "kube-api-access-n2s8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.060354 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2s8w\" (UniqueName: \"kubernetes.io/projected/2f5b8171-a997-49fe-ac00-eb90258f747a-kube-api-access-n2s8w\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.060390 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5b8171-a997-49fe-ac00-eb90258f747a-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.060403 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5b8171-a997-49fe-ac00-eb90258f747a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.264929 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2tsz" event={"ID":"2f5b8171-a997-49fe-ac00-eb90258f747a","Type":"ContainerDied","Data":"92ff5b34990be4d69c11fd4e786142bb24760a30af18ab9bf67e760c30b6f654"} Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.264986 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p2tsz" Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.264991 4931 scope.go:117] "RemoveContainer" containerID="69999e8efc16d246c63b1b1717e57fab2a6d6267ed8dfae22ce6e85640697570" Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.282424 4931 scope.go:117] "RemoveContainer" containerID="8d66ea5f5427eb66587c076b58fc47a9034ef5091990e6fc5f68d46257df6d06" Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.290595 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2tsz"] Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.294117 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2tsz"] Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.298710 4931 scope.go:117] "RemoveContainer" containerID="3d76e961881be63e5fb0b7aef98a8043913431da1a8852ed7504ffb2720e613e" Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.734699 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.735096 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.735287 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.736255 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.736611 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9" gracePeriod=600 Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.817506 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:03:23 crc kubenswrapper[4931]: I1129 00:03:23.817628 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.045276 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.045315 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.103363 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.165370 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.220957 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.221089 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.267282 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.272135 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9" exitCode=0 Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.272304 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9"} Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.304491 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.315224 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.429346 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.429637 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:03:24 crc kubenswrapper[4931]: I1129 00:03:24.479451 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:03:25 crc kubenswrapper[4931]: I1129 00:03:25.218513 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f5b8171-a997-49fe-ac00-eb90258f747a" path="/var/lib/kubelet/pods/2f5b8171-a997-49fe-ac00-eb90258f747a/volumes" Nov 29 00:03:25 crc kubenswrapper[4931]: I1129 00:03:25.279534 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"720e0a368eab3734412726774de172ffe278950e85b711a795977fe783122261"} Nov 29 00:03:25 crc kubenswrapper[4931]: I1129 00:03:25.320284 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:03:25 crc kubenswrapper[4931]: I1129 00:03:25.815627 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:03:25 crc kubenswrapper[4931]: I1129 00:03:25.815707 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:03:25 crc kubenswrapper[4931]: I1129 00:03:25.869877 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:03:26 crc kubenswrapper[4931]: I1129 00:03:26.323400 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:03:27 crc kubenswrapper[4931]: I1129 00:03:27.022660 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:03:27 crc kubenswrapper[4931]: I1129 00:03:27.022726 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:03:27 crc kubenswrapper[4931]: I1129 00:03:27.475693 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:03:27 crc kubenswrapper[4931]: I1129 00:03:27.475744 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:03:27 crc kubenswrapper[4931]: I1129 00:03:27.540710 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:03:27 crc kubenswrapper[4931]: I1129 00:03:27.851279 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6ngbj"] Nov 29 00:03:27 crc kubenswrapper[4931]: I1129 00:03:27.851611 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6ngbj" podUID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" containerName="registry-server" containerID="cri-o://a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3" gracePeriod=2 Nov 29 00:03:28 crc kubenswrapper[4931]: I1129 00:03:28.079771 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q9c9x" podUID="2e2b7be6-d4e7-4e92-b419-bda83df88970" containerName="registry-server" probeResult="failure" output=< Nov 29 00:03:28 crc kubenswrapper[4931]: timeout: failed to connect service ":50051" within 1s Nov 29 00:03:28 crc kubenswrapper[4931]: > Nov 29 00:03:28 crc kubenswrapper[4931]: I1129 00:03:28.348631 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.019662 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.156556 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26p65\" (UniqueName: \"kubernetes.io/projected/bf2c190f-062c-407a-87dd-48a5c4ac07d5-kube-api-access-26p65\") pod \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\" (UID: \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\") " Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.156638 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf2c190f-062c-407a-87dd-48a5c4ac07d5-catalog-content\") pod \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\" (UID: \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\") " Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.156716 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf2c190f-062c-407a-87dd-48a5c4ac07d5-utilities\") pod \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\" (UID: \"bf2c190f-062c-407a-87dd-48a5c4ac07d5\") " Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.157726 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf2c190f-062c-407a-87dd-48a5c4ac07d5-utilities" (OuterVolumeSpecName: "utilities") pod "bf2c190f-062c-407a-87dd-48a5c4ac07d5" (UID: "bf2c190f-062c-407a-87dd-48a5c4ac07d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.169954 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf2c190f-062c-407a-87dd-48a5c4ac07d5-kube-api-access-26p65" (OuterVolumeSpecName: "kube-api-access-26p65") pod "bf2c190f-062c-407a-87dd-48a5c4ac07d5" (UID: "bf2c190f-062c-407a-87dd-48a5c4ac07d5"). InnerVolumeSpecName "kube-api-access-26p65". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.205657 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf2c190f-062c-407a-87dd-48a5c4ac07d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf2c190f-062c-407a-87dd-48a5c4ac07d5" (UID: "bf2c190f-062c-407a-87dd-48a5c4ac07d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.245041 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g72fb"] Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.258438 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf2c190f-062c-407a-87dd-48a5c4ac07d5-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.258474 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26p65\" (UniqueName: \"kubernetes.io/projected/bf2c190f-062c-407a-87dd-48a5c4ac07d5-kube-api-access-26p65\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.258486 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf2c190f-062c-407a-87dd-48a5c4ac07d5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.308828 4931 generic.go:334] "Generic (PLEG): container finished" podID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" containerID="a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3" exitCode=0 Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.308903 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ngbj" event={"ID":"bf2c190f-062c-407a-87dd-48a5c4ac07d5","Type":"ContainerDied","Data":"a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3"} Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.308920 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6ngbj" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.308955 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6ngbj" event={"ID":"bf2c190f-062c-407a-87dd-48a5c4ac07d5","Type":"ContainerDied","Data":"da718e56cb1d2a7bc362178f791db351ef8e2b3f25f9b976a99fb47b6ceb096f"} Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.308978 4931 scope.go:117] "RemoveContainer" containerID="a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.309121 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g72fb" podUID="74999bec-b01b-4643-9cad-fa5e146089f5" containerName="registry-server" containerID="cri-o://b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74" gracePeriod=2 Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.324373 4931 scope.go:117] "RemoveContainer" containerID="9e85638a6b3e08156edc042510cb95e75810f327c1d8f01d51313e3f8e9a1893" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.336469 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6ngbj"] Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.340855 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6ngbj"] Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.342841 4931 scope.go:117] "RemoveContainer" containerID="3e00e2c6606ce8b1d26fb8ffa2ad0ba3ba2759475341049405ed4e67a5df3538" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.373986 4931 scope.go:117] "RemoveContainer" containerID="a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3" Nov 29 00:03:30 crc kubenswrapper[4931]: E1129 00:03:30.374315 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3\": container with ID starting with a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3 not found: ID does not exist" containerID="a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.374344 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3"} err="failed to get container status \"a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3\": rpc error: code = NotFound desc = could not find container \"a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3\": container with ID starting with a215162109a1a86e57a8fa4b726f074a26848763523248de0cf154d4e7fe66c3 not found: ID does not exist" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.374365 4931 scope.go:117] "RemoveContainer" containerID="9e85638a6b3e08156edc042510cb95e75810f327c1d8f01d51313e3f8e9a1893" Nov 29 00:03:30 crc kubenswrapper[4931]: E1129 00:03:30.374538 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e85638a6b3e08156edc042510cb95e75810f327c1d8f01d51313e3f8e9a1893\": container with ID starting with 9e85638a6b3e08156edc042510cb95e75810f327c1d8f01d51313e3f8e9a1893 not found: ID does not exist" containerID="9e85638a6b3e08156edc042510cb95e75810f327c1d8f01d51313e3f8e9a1893" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.374581 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e85638a6b3e08156edc042510cb95e75810f327c1d8f01d51313e3f8e9a1893"} err="failed to get container status \"9e85638a6b3e08156edc042510cb95e75810f327c1d8f01d51313e3f8e9a1893\": rpc error: code = NotFound desc = could not find container \"9e85638a6b3e08156edc042510cb95e75810f327c1d8f01d51313e3f8e9a1893\": container with ID starting with 9e85638a6b3e08156edc042510cb95e75810f327c1d8f01d51313e3f8e9a1893 not found: ID does not exist" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.374597 4931 scope.go:117] "RemoveContainer" containerID="3e00e2c6606ce8b1d26fb8ffa2ad0ba3ba2759475341049405ed4e67a5df3538" Nov 29 00:03:30 crc kubenswrapper[4931]: E1129 00:03:30.374842 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e00e2c6606ce8b1d26fb8ffa2ad0ba3ba2759475341049405ed4e67a5df3538\": container with ID starting with 3e00e2c6606ce8b1d26fb8ffa2ad0ba3ba2759475341049405ed4e67a5df3538 not found: ID does not exist" containerID="3e00e2c6606ce8b1d26fb8ffa2ad0ba3ba2759475341049405ed4e67a5df3538" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.374862 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e00e2c6606ce8b1d26fb8ffa2ad0ba3ba2759475341049405ed4e67a5df3538"} err="failed to get container status \"3e00e2c6606ce8b1d26fb8ffa2ad0ba3ba2759475341049405ed4e67a5df3538\": rpc error: code = NotFound desc = could not find container \"3e00e2c6606ce8b1d26fb8ffa2ad0ba3ba2759475341049405ed4e67a5df3538\": container with ID starting with 3e00e2c6606ce8b1d26fb8ffa2ad0ba3ba2759475341049405ed4e67a5df3538 not found: ID does not exist" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.666767 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.868148 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vmjf\" (UniqueName: \"kubernetes.io/projected/74999bec-b01b-4643-9cad-fa5e146089f5-kube-api-access-5vmjf\") pod \"74999bec-b01b-4643-9cad-fa5e146089f5\" (UID: \"74999bec-b01b-4643-9cad-fa5e146089f5\") " Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.868220 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74999bec-b01b-4643-9cad-fa5e146089f5-utilities\") pod \"74999bec-b01b-4643-9cad-fa5e146089f5\" (UID: \"74999bec-b01b-4643-9cad-fa5e146089f5\") " Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.868308 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74999bec-b01b-4643-9cad-fa5e146089f5-catalog-content\") pod \"74999bec-b01b-4643-9cad-fa5e146089f5\" (UID: \"74999bec-b01b-4643-9cad-fa5e146089f5\") " Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.869799 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74999bec-b01b-4643-9cad-fa5e146089f5-utilities" (OuterVolumeSpecName: "utilities") pod "74999bec-b01b-4643-9cad-fa5e146089f5" (UID: "74999bec-b01b-4643-9cad-fa5e146089f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.875184 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74999bec-b01b-4643-9cad-fa5e146089f5-kube-api-access-5vmjf" (OuterVolumeSpecName: "kube-api-access-5vmjf") pod "74999bec-b01b-4643-9cad-fa5e146089f5" (UID: "74999bec-b01b-4643-9cad-fa5e146089f5"). InnerVolumeSpecName "kube-api-access-5vmjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.970655 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vmjf\" (UniqueName: \"kubernetes.io/projected/74999bec-b01b-4643-9cad-fa5e146089f5-kube-api-access-5vmjf\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:30 crc kubenswrapper[4931]: I1129 00:03:30.970701 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74999bec-b01b-4643-9cad-fa5e146089f5-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.037628 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74999bec-b01b-4643-9cad-fa5e146089f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74999bec-b01b-4643-9cad-fa5e146089f5" (UID: "74999bec-b01b-4643-9cad-fa5e146089f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.071414 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74999bec-b01b-4643-9cad-fa5e146089f5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.222060 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" path="/var/lib/kubelet/pods/bf2c190f-062c-407a-87dd-48a5c4ac07d5/volumes" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.332964 4931 generic.go:334] "Generic (PLEG): container finished" podID="74999bec-b01b-4643-9cad-fa5e146089f5" containerID="b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74" exitCode=0 Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.333006 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g72fb" event={"ID":"74999bec-b01b-4643-9cad-fa5e146089f5","Type":"ContainerDied","Data":"b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74"} Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.333028 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g72fb" event={"ID":"74999bec-b01b-4643-9cad-fa5e146089f5","Type":"ContainerDied","Data":"21b11a34f36e3b7223945b2e07bc68750c563f5052dc415aba1f4a4b762f3d1f"} Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.333044 4931 scope.go:117] "RemoveContainer" containerID="b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.333160 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g72fb" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.351019 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g72fb"] Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.353610 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g72fb"] Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.353714 4931 scope.go:117] "RemoveContainer" containerID="522829861b2063315b77d619f9c3c5482f97715d07bc8fbbf3c6a2014a5d708a" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.367677 4931 scope.go:117] "RemoveContainer" containerID="249487ee69bf04eccbc907c850cfcbba1b496887eee5a30df13961dc389d8fa3" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.393365 4931 scope.go:117] "RemoveContainer" containerID="b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74" Nov 29 00:03:31 crc kubenswrapper[4931]: E1129 00:03:31.393695 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74\": container with ID starting with b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74 not found: ID does not exist" containerID="b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.393729 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74"} err="failed to get container status \"b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74\": rpc error: code = NotFound desc = could not find container \"b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74\": container with ID starting with b11b63e12b6718e2ea09fdee1fd334f4a67c8a77b1119d847e472274e25d2d74 not found: ID does not exist" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.393749 4931 scope.go:117] "RemoveContainer" containerID="522829861b2063315b77d619f9c3c5482f97715d07bc8fbbf3c6a2014a5d708a" Nov 29 00:03:31 crc kubenswrapper[4931]: E1129 00:03:31.394531 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"522829861b2063315b77d619f9c3c5482f97715d07bc8fbbf3c6a2014a5d708a\": container with ID starting with 522829861b2063315b77d619f9c3c5482f97715d07bc8fbbf3c6a2014a5d708a not found: ID does not exist" containerID="522829861b2063315b77d619f9c3c5482f97715d07bc8fbbf3c6a2014a5d708a" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.394554 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"522829861b2063315b77d619f9c3c5482f97715d07bc8fbbf3c6a2014a5d708a"} err="failed to get container status \"522829861b2063315b77d619f9c3c5482f97715d07bc8fbbf3c6a2014a5d708a\": rpc error: code = NotFound desc = could not find container \"522829861b2063315b77d619f9c3c5482f97715d07bc8fbbf3c6a2014a5d708a\": container with ID starting with 522829861b2063315b77d619f9c3c5482f97715d07bc8fbbf3c6a2014a5d708a not found: ID does not exist" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.394565 4931 scope.go:117] "RemoveContainer" containerID="249487ee69bf04eccbc907c850cfcbba1b496887eee5a30df13961dc389d8fa3" Nov 29 00:03:31 crc kubenswrapper[4931]: E1129 00:03:31.394795 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"249487ee69bf04eccbc907c850cfcbba1b496887eee5a30df13961dc389d8fa3\": container with ID starting with 249487ee69bf04eccbc907c850cfcbba1b496887eee5a30df13961dc389d8fa3 not found: ID does not exist" containerID="249487ee69bf04eccbc907c850cfcbba1b496887eee5a30df13961dc389d8fa3" Nov 29 00:03:31 crc kubenswrapper[4931]: I1129 00:03:31.394829 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"249487ee69bf04eccbc907c850cfcbba1b496887eee5a30df13961dc389d8fa3"} err="failed to get container status \"249487ee69bf04eccbc907c850cfcbba1b496887eee5a30df13961dc389d8fa3\": rpc error: code = NotFound desc = could not find container \"249487ee69bf04eccbc907c850cfcbba1b496887eee5a30df13961dc389d8fa3\": container with ID starting with 249487ee69bf04eccbc907c850cfcbba1b496887eee5a30df13961dc389d8fa3 not found: ID does not exist" Nov 29 00:03:33 crc kubenswrapper[4931]: I1129 00:03:33.224037 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74999bec-b01b-4643-9cad-fa5e146089f5" path="/var/lib/kubelet/pods/74999bec-b01b-4643-9cad-fa5e146089f5/volumes" Nov 29 00:03:34 crc kubenswrapper[4931]: I1129 00:03:34.268144 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.044268 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cdzck"] Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.044626 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cdzck" podUID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" containerName="registry-server" containerID="cri-o://e94a1ee77ce918558bc0b088d0a891a5e2eec1ffabdf7c7396cec64f9b099925" gracePeriod=2 Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.372197 4931 generic.go:334] "Generic (PLEG): container finished" podID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" containerID="e94a1ee77ce918558bc0b088d0a891a5e2eec1ffabdf7c7396cec64f9b099925" exitCode=0 Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.372273 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cdzck" event={"ID":"3e56db6c-4dd6-461b-8f37-5ce71ea285b3","Type":"ContainerDied","Data":"e94a1ee77ce918558bc0b088d0a891a5e2eec1ffabdf7c7396cec64f9b099925"} Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.443931 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.630392 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d92sp\" (UniqueName: \"kubernetes.io/projected/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-kube-api-access-d92sp\") pod \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\" (UID: \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\") " Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.630475 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-catalog-content\") pod \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\" (UID: \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\") " Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.630561 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-utilities\") pod \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\" (UID: \"3e56db6c-4dd6-461b-8f37-5ce71ea285b3\") " Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.631425 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-utilities" (OuterVolumeSpecName: "utilities") pod "3e56db6c-4dd6-461b-8f37-5ce71ea285b3" (UID: "3e56db6c-4dd6-461b-8f37-5ce71ea285b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.640442 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-kube-api-access-d92sp" (OuterVolumeSpecName: "kube-api-access-d92sp") pod "3e56db6c-4dd6-461b-8f37-5ce71ea285b3" (UID: "3e56db6c-4dd6-461b-8f37-5ce71ea285b3"). InnerVolumeSpecName "kube-api-access-d92sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.713101 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e56db6c-4dd6-461b-8f37-5ce71ea285b3" (UID: "3e56db6c-4dd6-461b-8f37-5ce71ea285b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.732258 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.732294 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d92sp\" (UniqueName: \"kubernetes.io/projected/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-kube-api-access-d92sp\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.732305 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e56db6c-4dd6-461b-8f37-5ce71ea285b3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:35 crc kubenswrapper[4931]: I1129 00:03:35.831156 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kshsm"] Nov 29 00:03:36 crc kubenswrapper[4931]: I1129 00:03:36.381351 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cdzck" event={"ID":"3e56db6c-4dd6-461b-8f37-5ce71ea285b3","Type":"ContainerDied","Data":"9b06a9337ef5b7ec20ec48dd14448c6e3769b1a25eaf3797923ea2fa352e8810"} Nov 29 00:03:36 crc kubenswrapper[4931]: I1129 00:03:36.381699 4931 scope.go:117] "RemoveContainer" containerID="e94a1ee77ce918558bc0b088d0a891a5e2eec1ffabdf7c7396cec64f9b099925" Nov 29 00:03:36 crc kubenswrapper[4931]: I1129 00:03:36.381906 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cdzck" Nov 29 00:03:36 crc kubenswrapper[4931]: I1129 00:03:36.403506 4931 scope.go:117] "RemoveContainer" containerID="5ad8f53a6c80d568856ef481f45a9b2779cb337002ccf1987f8625e079519d54" Nov 29 00:03:36 crc kubenswrapper[4931]: I1129 00:03:36.424123 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cdzck"] Nov 29 00:03:36 crc kubenswrapper[4931]: I1129 00:03:36.433233 4931 scope.go:117] "RemoveContainer" containerID="e986bdc9e9e692b98b22b797ae3c873f22b4cab965633438f596d3bea113c0d9" Nov 29 00:03:36 crc kubenswrapper[4931]: I1129 00:03:36.435138 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cdzck"] Nov 29 00:03:37 crc kubenswrapper[4931]: I1129 00:03:37.072963 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:03:37 crc kubenswrapper[4931]: I1129 00:03:37.115555 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:03:37 crc kubenswrapper[4931]: I1129 00:03:37.219661 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" path="/var/lib/kubelet/pods/3e56db6c-4dd6-461b-8f37-5ce71ea285b3/volumes" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.798860 4931 file.go:109] "Unable to process watch event" err="can't process config file \"/etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml\": /etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml: couldn't parse as pod(Object 'Kind' is missing in 'null'), please check config file" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801303 4931 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801540 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f5b8171-a997-49fe-ac00-eb90258f747a" containerName="registry-server" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801557 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f5b8171-a997-49fe-ac00-eb90258f747a" containerName="registry-server" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801569 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" containerName="extract-utilities" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801575 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" containerName="extract-utilities" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801586 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f5b8171-a997-49fe-ac00-eb90258f747a" containerName="extract-content" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801592 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f5b8171-a997-49fe-ac00-eb90258f747a" containerName="extract-content" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801600 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" containerName="extract-utilities" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801607 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" containerName="extract-utilities" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801614 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" containerName="registry-server" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801620 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" containerName="registry-server" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801630 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74999bec-b01b-4643-9cad-fa5e146089f5" containerName="extract-content" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801637 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="74999bec-b01b-4643-9cad-fa5e146089f5" containerName="extract-content" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801647 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" containerName="extract-content" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801653 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" containerName="extract-content" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801661 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" containerName="extract-content" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801666 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" containerName="extract-content" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801672 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" containerName="registry-server" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801677 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" containerName="registry-server" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801686 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74999bec-b01b-4643-9cad-fa5e146089f5" containerName="registry-server" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801691 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="74999bec-b01b-4643-9cad-fa5e146089f5" containerName="registry-server" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801698 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f5b8171-a997-49fe-ac00-eb90258f747a" containerName="extract-utilities" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801703 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f5b8171-a997-49fe-ac00-eb90258f747a" containerName="extract-utilities" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801712 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74999bec-b01b-4643-9cad-fa5e146089f5" containerName="extract-utilities" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801717 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="74999bec-b01b-4643-9cad-fa5e146089f5" containerName="extract-utilities" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.801725 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e67aff2-bf6c-4f67-922a-c0454e7026ff" containerName="pruner" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801731 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e67aff2-bf6c-4f67-922a-c0454e7026ff" containerName="pruner" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801842 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf2c190f-062c-407a-87dd-48a5c4ac07d5" containerName="registry-server" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801855 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e67aff2-bf6c-4f67-922a-c0454e7026ff" containerName="pruner" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801862 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e56db6c-4dd6-461b-8f37-5ce71ea285b3" containerName="registry-server" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801870 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f5b8171-a997-49fe-ac00-eb90258f747a" containerName="registry-server" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.801879 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="74999bec-b01b-4643-9cad-fa5e146089f5" containerName="registry-server" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.802261 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.802681 4931 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803132 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f" gracePeriod=15 Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803187 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c" gracePeriod=15 Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803286 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8" gracePeriod=15 Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803339 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082" gracePeriod=15 Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803411 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975" gracePeriod=15 Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803709 4931 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.803843 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803853 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.803863 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803869 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.803876 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803882 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.803894 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803901 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.803909 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803915 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.803923 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803929 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 29 00:03:52 crc kubenswrapper[4931]: E1129 00:03:52.803936 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.803942 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.804021 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.804030 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.804037 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.804044 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.804053 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.804061 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.989925 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.990340 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.990399 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.990492 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.990552 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.990579 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.990607 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:03:52 crc kubenswrapper[4931]: I1129 00:03:52.990641 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.091715 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.091842 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.091911 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.091924 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.092014 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.091949 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.092116 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.092132 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.092035 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.092208 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.092254 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.092303 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.092350 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.092517 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.092496 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.092616 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.487113 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.490559 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.491635 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f" exitCode=0 Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.491674 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8" exitCode=0 Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.491689 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c" exitCode=0 Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.491702 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082" exitCode=2 Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.491712 4931 scope.go:117] "RemoveContainer" containerID="7fabd739cda843cb77aecbf7fe2fcf3d87e6d9b99f40bbb549997c59dfffe70e" Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.495043 4931 generic.go:334] "Generic (PLEG): container finished" podID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" containerID="b36f834a1c60b2b466fb494d417c6565be12b8ac9d07e89bd8693fea59009422" exitCode=0 Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.495090 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"4b0e513e-266b-4aeb-9c04-e7b07d28ec19","Type":"ContainerDied","Data":"b36f834a1c60b2b466fb494d417c6565be12b8ac9d07e89bd8693fea59009422"} Nov 29 00:03:53 crc kubenswrapper[4931]: I1129 00:03:53.497580 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:54 crc kubenswrapper[4931]: I1129 00:03:54.506234 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 29 00:03:54 crc kubenswrapper[4931]: I1129 00:03:54.843999 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:54 crc kubenswrapper[4931]: I1129 00:03:54.845008 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.039168 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-kubelet-dir\") pod \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\" (UID: \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\") " Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.039459 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-var-lock\") pod \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\" (UID: \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\") " Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.039317 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4b0e513e-266b-4aeb-9c04-e7b07d28ec19" (UID: "4b0e513e-266b-4aeb-9c04-e7b07d28ec19"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.039502 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-kube-api-access\") pod \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\" (UID: \"4b0e513e-266b-4aeb-9c04-e7b07d28ec19\") " Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.039551 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-var-lock" (OuterVolumeSpecName: "var-lock") pod "4b0e513e-266b-4aeb-9c04-e7b07d28ec19" (UID: "4b0e513e-266b-4aeb-9c04-e7b07d28ec19"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.039735 4931 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.039747 4931 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-var-lock\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.044951 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4b0e513e-266b-4aeb-9c04-e7b07d28ec19" (UID: "4b0e513e-266b-4aeb-9c04-e7b07d28ec19"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.141039 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b0e513e-266b-4aeb-9c04-e7b07d28ec19-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.397643 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.400034 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.401024 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.401504 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.444650 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.444750 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.444840 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.444871 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.444952 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.444996 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.445283 4931 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.445310 4931 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.445327 4931 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.518920 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.520389 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975" exitCode=0 Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.520520 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.520525 4931 scope.go:117] "RemoveContainer" containerID="8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.523699 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"4b0e513e-266b-4aeb-9c04-e7b07d28ec19","Type":"ContainerDied","Data":"44f0898bd30a954dd7b54375f20822ecafc4eb76d8106d4f9cc7fce3682eddde"} Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.523756 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44f0898bd30a954dd7b54375f20822ecafc4eb76d8106d4f9cc7fce3682eddde" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.523867 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.540467 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.540894 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.547630 4931 scope.go:117] "RemoveContainer" containerID="3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.554387 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.555041 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.569960 4931 scope.go:117] "RemoveContainer" containerID="a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.589239 4931 scope.go:117] "RemoveContainer" containerID="8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.608229 4931 scope.go:117] "RemoveContainer" containerID="cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.630517 4931 scope.go:117] "RemoveContainer" containerID="23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.660086 4931 scope.go:117] "RemoveContainer" containerID="8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f" Nov 29 00:03:55 crc kubenswrapper[4931]: E1129 00:03:55.660622 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\": container with ID starting with 8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f not found: ID does not exist" containerID="8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.660697 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f"} err="failed to get container status \"8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\": rpc error: code = NotFound desc = could not find container \"8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f\": container with ID starting with 8e2ddaa71774bbe6dcbbc1935c8f3f6b1c68079a9c087190f96e88d57ff9134f not found: ID does not exist" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.660735 4931 scope.go:117] "RemoveContainer" containerID="3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8" Nov 29 00:03:55 crc kubenswrapper[4931]: E1129 00:03:55.661227 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\": container with ID starting with 3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8 not found: ID does not exist" containerID="3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.661292 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8"} err="failed to get container status \"3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\": rpc error: code = NotFound desc = could not find container \"3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8\": container with ID starting with 3a1fb640454c9bc6dc73be314cc79d60b29eee0b5f6fd58753e6d56b7e6731f8 not found: ID does not exist" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.661345 4931 scope.go:117] "RemoveContainer" containerID="a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c" Nov 29 00:03:55 crc kubenswrapper[4931]: E1129 00:03:55.661770 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\": container with ID starting with a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c not found: ID does not exist" containerID="a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.661840 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c"} err="failed to get container status \"a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\": rpc error: code = NotFound desc = could not find container \"a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c\": container with ID starting with a061a1c23a96eb3cc185b3a72ac6c60ac566fdb5bf77f5cd3f3b24bc4cbfa35c not found: ID does not exist" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.661869 4931 scope.go:117] "RemoveContainer" containerID="8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082" Nov 29 00:03:55 crc kubenswrapper[4931]: E1129 00:03:55.662384 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\": container with ID starting with 8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082 not found: ID does not exist" containerID="8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.662448 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082"} err="failed to get container status \"8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\": rpc error: code = NotFound desc = could not find container \"8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082\": container with ID starting with 8b630251d65e91b80190f3fe41f1f9ee1f5e88ce37389ead83fb4d025d2a3082 not found: ID does not exist" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.662483 4931 scope.go:117] "RemoveContainer" containerID="cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975" Nov 29 00:03:55 crc kubenswrapper[4931]: E1129 00:03:55.662853 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\": container with ID starting with cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975 not found: ID does not exist" containerID="cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.662894 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975"} err="failed to get container status \"cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\": rpc error: code = NotFound desc = could not find container \"cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975\": container with ID starting with cb905b574f7e474b2e03f163f77fca143586fd589542353dfa23853a609cf975 not found: ID does not exist" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.662918 4931 scope.go:117] "RemoveContainer" containerID="23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21" Nov 29 00:03:55 crc kubenswrapper[4931]: E1129 00:03:55.663446 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\": container with ID starting with 23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21 not found: ID does not exist" containerID="23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21" Nov 29 00:03:55 crc kubenswrapper[4931]: I1129 00:03:55.663483 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21"} err="failed to get container status \"23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\": rpc error: code = NotFound desc = could not find container \"23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21\": container with ID starting with 23be973296b269f9ee5b46e1828da8a78035a7bcad04321418d2490e0ca68a21 not found: ID does not exist" Nov 29 00:03:56 crc kubenswrapper[4931]: E1129 00:03:56.079151 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:03:56Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:03:56Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:03:56Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:03:56Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:56 crc kubenswrapper[4931]: E1129 00:03:56.079659 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:56 crc kubenswrapper[4931]: E1129 00:03:56.080143 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:56 crc kubenswrapper[4931]: E1129 00:03:56.081011 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:56 crc kubenswrapper[4931]: E1129 00:03:56.081683 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:56 crc kubenswrapper[4931]: E1129 00:03:56.081785 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 00:03:57 crc kubenswrapper[4931]: I1129 00:03:57.216519 4931 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:57 crc kubenswrapper[4931]: I1129 00:03:57.217606 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:57 crc kubenswrapper[4931]: I1129 00:03:57.222273 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 29 00:03:57 crc kubenswrapper[4931]: E1129 00:03:57.831439 4931 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:57 crc kubenswrapper[4931]: I1129 00:03:57.831949 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:57 crc kubenswrapper[4931]: W1129 00:03:57.864654 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-c1e2440ec16df8a22803ff88a5ada7aaac58cdd5bb84e7e83f86736aca5be5c5 WatchSource:0}: Error finding container c1e2440ec16df8a22803ff88a5ada7aaac58cdd5bb84e7e83f86736aca5be5c5: Status 404 returned error can't find the container with id c1e2440ec16df8a22803ff88a5ada7aaac58cdd5bb84e7e83f86736aca5be5c5 Nov 29 00:03:57 crc kubenswrapper[4931]: E1129 00:03:57.869432 4931 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.18:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187c516294d803fd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-29 00:03:57.868221437 +0000 UTC m=+251.030114669,LastTimestamp:2025-11-29 00:03:57.868221437 +0000 UTC m=+251.030114669,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 29 00:03:58 crc kubenswrapper[4931]: I1129 00:03:58.550752 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"31e9ae5daa0e45e16ccea59f5bd0d50977658114846a3705f80646ac07a1c777"} Nov 29 00:03:58 crc kubenswrapper[4931]: I1129 00:03:58.551081 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c1e2440ec16df8a22803ff88a5ada7aaac58cdd5bb84e7e83f86736aca5be5c5"} Nov 29 00:03:58 crc kubenswrapper[4931]: I1129 00:03:58.551684 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:03:58 crc kubenswrapper[4931]: E1129 00:03:58.551787 4931 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:03:59 crc kubenswrapper[4931]: E1129 00:03:59.546390 4931 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.18:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187c516294d803fd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-29 00:03:57.868221437 +0000 UTC m=+251.030114669,LastTimestamp:2025-11-29 00:03:57.868221437 +0000 UTC m=+251.030114669,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 29 00:04:00 crc kubenswrapper[4931]: E1129 00:04:00.790215 4931 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:00 crc kubenswrapper[4931]: E1129 00:04:00.790737 4931 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:00 crc kubenswrapper[4931]: E1129 00:04:00.791060 4931 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:00 crc kubenswrapper[4931]: E1129 00:04:00.791243 4931 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:00 crc kubenswrapper[4931]: E1129 00:04:00.791421 4931 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:00 crc kubenswrapper[4931]: I1129 00:04:00.791443 4931 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 29 00:04:00 crc kubenswrapper[4931]: E1129 00:04:00.791625 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="200ms" Nov 29 00:04:00 crc kubenswrapper[4931]: I1129 00:04:00.860436 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" containerName="oauth-openshift" containerID="cri-o://dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f" gracePeriod=15 Nov 29 00:04:00 crc kubenswrapper[4931]: E1129 00:04:00.992476 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="400ms" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.293339 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.294260 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.294729 4931 status_manager.go:851] "Failed to get status for pod" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-kshsm\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.333822 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-router-certs\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.333872 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tw99\" (UniqueName: \"kubernetes.io/projected/199fba37-4269-4fb2-9249-b384fc2f2905-kube-api-access-6tw99\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.333897 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-audit-policies\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.333920 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-serving-cert\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.333950 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/199fba37-4269-4fb2-9249-b384fc2f2905-audit-dir\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.333986 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-error\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.334006 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-session\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.334033 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-cliconfig\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.334056 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-trusted-ca-bundle\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.334082 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-idp-0-file-data\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.334102 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-login\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.334134 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-ocp-branding-template\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.334159 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-service-ca\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.334200 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-provider-selection\") pod \"199fba37-4269-4fb2-9249-b384fc2f2905\" (UID: \"199fba37-4269-4fb2-9249-b384fc2f2905\") " Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.335279 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/199fba37-4269-4fb2-9249-b384fc2f2905-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.335359 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.335466 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.336070 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.337102 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.341710 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.342297 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.343664 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.343784 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.344107 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.344127 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/199fba37-4269-4fb2-9249-b384fc2f2905-kube-api-access-6tw99" (OuterVolumeSpecName: "kube-api-access-6tw99") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "kube-api-access-6tw99". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.346671 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.347498 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.347872 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "199fba37-4269-4fb2-9249-b384fc2f2905" (UID: "199fba37-4269-4fb2-9249-b384fc2f2905"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:04:01 crc kubenswrapper[4931]: E1129 00:04:01.393584 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="800ms" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.435681 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.435755 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.435780 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.435849 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.435881 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.435907 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.435927 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tw99\" (UniqueName: \"kubernetes.io/projected/199fba37-4269-4fb2-9249-b384fc2f2905-kube-api-access-6tw99\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.435950 4931 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.435971 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.435994 4931 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/199fba37-4269-4fb2-9249-b384fc2f2905-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.436014 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.436033 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.436053 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.436073 4931 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/199fba37-4269-4fb2-9249-b384fc2f2905-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.570661 4931 generic.go:334] "Generic (PLEG): container finished" podID="199fba37-4269-4fb2-9249-b384fc2f2905" containerID="dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f" exitCode=0 Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.570782 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" event={"ID":"199fba37-4269-4fb2-9249-b384fc2f2905","Type":"ContainerDied","Data":"dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f"} Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.570801 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.570889 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" event={"ID":"199fba37-4269-4fb2-9249-b384fc2f2905","Type":"ContainerDied","Data":"a50bcd352c4e32faa2853bccdcdb5544622ed59f9ed0bd0e442be657d803c135"} Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.570948 4931 scope.go:117] "RemoveContainer" containerID="dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.571791 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.572399 4931 status_manager.go:851] "Failed to get status for pod" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-kshsm\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.598928 4931 scope.go:117] "RemoveContainer" containerID="dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f" Nov 29 00:04:01 crc kubenswrapper[4931]: E1129 00:04:01.599691 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f\": container with ID starting with dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f not found: ID does not exist" containerID="dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.599954 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f"} err="failed to get container status \"dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f\": rpc error: code = NotFound desc = could not find container \"dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f\": container with ID starting with dac099a2cd01fb15412aa8723afc1944c4f441badcd89d2e3fb468a13e9e414f not found: ID does not exist" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.608218 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:01 crc kubenswrapper[4931]: I1129 00:04:01.608763 4931 status_manager.go:851] "Failed to get status for pod" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-kshsm\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:02 crc kubenswrapper[4931]: E1129 00:04:02.195150 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="1.6s" Nov 29 00:04:03 crc kubenswrapper[4931]: E1129 00:04:03.796455 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="3.2s" Nov 29 00:04:05 crc kubenswrapper[4931]: E1129 00:04:05.283210 4931 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" volumeName="registry-storage" Nov 29 00:04:06 crc kubenswrapper[4931]: E1129 00:04:06.465869 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:04:06Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:04:06Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:04:06Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-29T00:04:06Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:06 crc kubenswrapper[4931]: E1129 00:04:06.466438 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:06 crc kubenswrapper[4931]: E1129 00:04:06.466900 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:06 crc kubenswrapper[4931]: E1129 00:04:06.467616 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:06 crc kubenswrapper[4931]: E1129 00:04:06.468095 4931 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:06 crc kubenswrapper[4931]: E1129 00:04:06.468139 4931 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 29 00:04:06 crc kubenswrapper[4931]: E1129 00:04:06.997576 4931 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="6.4s" Nov 29 00:04:07 crc kubenswrapper[4931]: I1129 00:04:07.206703 4931 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 29 00:04:07 crc kubenswrapper[4931]: I1129 00:04:07.206786 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 29 00:04:07 crc kubenswrapper[4931]: I1129 00:04:07.217139 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:07 crc kubenswrapper[4931]: I1129 00:04:07.217494 4931 status_manager.go:851] "Failed to get status for pod" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-kshsm\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:07 crc kubenswrapper[4931]: I1129 00:04:07.617625 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 29 00:04:07 crc kubenswrapper[4931]: I1129 00:04:07.617732 4931 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c" exitCode=1 Nov 29 00:04:07 crc kubenswrapper[4931]: I1129 00:04:07.617788 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c"} Nov 29 00:04:07 crc kubenswrapper[4931]: I1129 00:04:07.618614 4931 scope.go:117] "RemoveContainer" containerID="8766b0e2bf2fd6225239efef068905762a37355289ff13bbd09950513be4f62c" Nov 29 00:04:07 crc kubenswrapper[4931]: I1129 00:04:07.619008 4931 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:07 crc kubenswrapper[4931]: I1129 00:04:07.619670 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:07 crc kubenswrapper[4931]: I1129 00:04:07.620324 4931 status_manager.go:851] "Failed to get status for pod" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-kshsm\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.211277 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.212066 4931 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.212462 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.213024 4931 status_manager.go:851] "Failed to get status for pod" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-kshsm\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.227664 4931 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.227709 4931 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb" Nov 29 00:04:08 crc kubenswrapper[4931]: E1129 00:04:08.228338 4931 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.228736 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:04:08 crc kubenswrapper[4931]: W1129 00:04:08.250839 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-9f3734ba17128fa568a92b8e5394f019b17b03d3fdf932817d8d6fbad0cbe44b WatchSource:0}: Error finding container 9f3734ba17128fa568a92b8e5394f019b17b03d3fdf932817d8d6fbad0cbe44b: Status 404 returned error can't find the container with id 9f3734ba17128fa568a92b8e5394f019b17b03d3fdf932817d8d6fbad0cbe44b Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.629970 4931 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="004356d2ea384a75da7171555834c7538b4e51eec6086739b7167e0816f4fe69" exitCode=0 Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.630065 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"004356d2ea384a75da7171555834c7538b4e51eec6086739b7167e0816f4fe69"} Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.630153 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9f3734ba17128fa568a92b8e5394f019b17b03d3fdf932817d8d6fbad0cbe44b"} Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.630595 4931 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.630631 4931 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb" Nov 29 00:04:08 crc kubenswrapper[4931]: E1129 00:04:08.631307 4931 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.631372 4931 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.631922 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.632557 4931 status_manager.go:851] "Failed to get status for pod" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-kshsm\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.635531 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.635633 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4035b61bb8c6c61c12db24e399cf278c74d2c5a3578891fba7792a41864461d3"} Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.636696 4931 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.637239 4931 status_manager.go:851] "Failed to get status for pod" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:08 crc kubenswrapper[4931]: I1129 00:04:08.637794 4931 status_manager.go:851] "Failed to get status for pod" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" pod="openshift-authentication/oauth-openshift-558db77b4-kshsm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-kshsm\": dial tcp 38.102.83.18:6443: connect: connection refused" Nov 29 00:04:09 crc kubenswrapper[4931]: I1129 00:04:09.653001 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"33a95db58cc44564c3f22e1a4811513a13dd17e03d14bf3da4ff569694a4d985"} Nov 29 00:04:09 crc kubenswrapper[4931]: I1129 00:04:09.653285 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2d4e94e62135601840cf36660da955f4677ddbd8af103c90af18c0dc3d900abf"} Nov 29 00:04:09 crc kubenswrapper[4931]: I1129 00:04:09.653296 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b48d5b270664b53a58587b16ffb598e6ae5e515b3ee56e76475adb46eb37c6fb"} Nov 29 00:04:10 crc kubenswrapper[4931]: I1129 00:04:10.661942 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0c3bea7606a83be621af0e730e84f4c54280dae0517ad12c64e656ebd08df6e0"} Nov 29 00:04:10 crc kubenswrapper[4931]: I1129 00:04:10.662268 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"32abbb54fc531b82dca40ae5926f9aa2d7fa6d26577f22f3d6e35f3b93f5f966"} Nov 29 00:04:10 crc kubenswrapper[4931]: I1129 00:04:10.662378 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:04:10 crc kubenswrapper[4931]: I1129 00:04:10.662484 4931 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb" Nov 29 00:04:10 crc kubenswrapper[4931]: I1129 00:04:10.662511 4931 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb" Nov 29 00:04:11 crc kubenswrapper[4931]: I1129 00:04:11.657585 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 00:04:11 crc kubenswrapper[4931]: I1129 00:04:11.662042 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 00:04:11 crc kubenswrapper[4931]: I1129 00:04:11.666623 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 00:04:13 crc kubenswrapper[4931]: I1129 00:04:13.229911 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:04:13 crc kubenswrapper[4931]: I1129 00:04:13.230345 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:04:13 crc kubenswrapper[4931]: I1129 00:04:13.239518 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:04:15 crc kubenswrapper[4931]: I1129 00:04:15.691238 4931 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:04:16 crc kubenswrapper[4931]: I1129 00:04:16.697925 4931 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb" Nov 29 00:04:16 crc kubenswrapper[4931]: I1129 00:04:16.697977 4931 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb" Nov 29 00:04:16 crc kubenswrapper[4931]: I1129 00:04:16.706328 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:04:17 crc kubenswrapper[4931]: I1129 00:04:17.224369 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 29 00:04:17 crc kubenswrapper[4931]: I1129 00:04:17.235224 4931 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="86bb6ccf-4a63-4dc2-b863-6d8e4691e188" Nov 29 00:04:17 crc kubenswrapper[4931]: I1129 00:04:17.704170 4931 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb" Nov 29 00:04:17 crc kubenswrapper[4931]: I1129 00:04:17.704228 4931 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb" Nov 29 00:04:17 crc kubenswrapper[4931]: I1129 00:04:17.707047 4931 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="86bb6ccf-4a63-4dc2-b863-6d8e4691e188" Nov 29 00:04:25 crc kubenswrapper[4931]: I1129 00:04:25.743084 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 29 00:04:25 crc kubenswrapper[4931]: I1129 00:04:25.873975 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 29 00:04:26 crc kubenswrapper[4931]: I1129 00:04:26.426798 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 29 00:04:26 crc kubenswrapper[4931]: I1129 00:04:26.713313 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 29 00:04:26 crc kubenswrapper[4931]: I1129 00:04:26.818183 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 29 00:04:27 crc kubenswrapper[4931]: I1129 00:04:27.128506 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 29 00:04:27 crc kubenswrapper[4931]: I1129 00:04:27.309609 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 29 00:04:27 crc kubenswrapper[4931]: I1129 00:04:27.364637 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 29 00:04:27 crc kubenswrapper[4931]: I1129 00:04:27.524325 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 29 00:04:27 crc kubenswrapper[4931]: I1129 00:04:27.623710 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 29 00:04:27 crc kubenswrapper[4931]: I1129 00:04:27.676563 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 29 00:04:27 crc kubenswrapper[4931]: I1129 00:04:27.707259 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 29 00:04:27 crc kubenswrapper[4931]: I1129 00:04:27.901856 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.076029 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.179229 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.317033 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.361288 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.383645 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.531326 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.610831 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.659116 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.735575 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.783514 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.885537 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.888326 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.972435 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 29 00:04:28 crc kubenswrapper[4931]: I1129 00:04:28.975260 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.250082 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.291795 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.293551 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.296451 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.495125 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.498225 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.500463 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.523697 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.530700 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.549238 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.558977 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.629526 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.632164 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.642970 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.713600 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.716852 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 29 00:04:29 crc kubenswrapper[4931]: I1129 00:04:29.801633 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.027385 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.038616 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.046825 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.068415 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.245889 4931 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.260367 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.293928 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.429356 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.445985 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.587336 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.851835 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.890152 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.911982 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.924451 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.953555 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.964887 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.965108 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.979184 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 29 00:04:30 crc kubenswrapper[4931]: I1129 00:04:30.983538 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.019069 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.030861 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.043301 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.186551 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.196397 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.228161 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.315548 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.428558 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.439986 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.620171 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.656567 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.700680 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.733595 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.734120 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.807671 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.828983 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.834913 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.904441 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 29 00:04:31 crc kubenswrapper[4931]: I1129 00:04:31.977309 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.029402 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.184184 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.254175 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.290118 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.317947 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.325331 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.370888 4931 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.434623 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.434757 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.447311 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.461802 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.471140 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.486733 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.510197 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.576342 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.632163 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.646552 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.857084 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.901731 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.925910 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 29 00:04:32 crc kubenswrapper[4931]: I1129 00:04:32.964853 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.093003 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.210689 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.217497 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.315948 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.331850 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.373494 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.509065 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.547560 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.637932 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.673006 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.702626 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.728400 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.804690 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.840879 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.880149 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.914596 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.981896 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 29 00:04:33 crc kubenswrapper[4931]: I1129 00:04:33.996675 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.075376 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.082032 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.162532 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.358522 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.360971 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.365955 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.438928 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.495469 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.500085 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.552744 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.661795 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.705474 4931 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.710476 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-kshsm"] Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.710534 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-57569d6b9d-stjx8","openshift-kube-apiserver/kube-apiserver-crc"] Nov 29 00:04:34 crc kubenswrapper[4931]: E1129 00:04:34.710775 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" containerName="installer" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.710827 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" containerName="installer" Nov 29 00:04:34 crc kubenswrapper[4931]: E1129 00:04:34.710849 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" containerName="oauth-openshift" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.710864 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" containerName="oauth-openshift" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.711141 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" containerName="oauth-openshift" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.711176 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b0e513e-266b-4aeb-9c04-e7b07d28ec19" containerName="installer" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.711191 4931 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.711224 4931 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0bb9b2dd-2c38-4689-97f9-1b8cf4f039cb" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.711593 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.717014 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.717329 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.717546 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.717647 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.717769 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.718195 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.718524 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.718591 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.718672 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.718540 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.719352 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.720200 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.720314 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.730414 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.730428 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.733048 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.741046 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.741028787 podStartE2EDuration="19.741028787s" podCreationTimestamp="2025-11-29 00:04:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:04:34.738715357 +0000 UTC m=+287.900608619" watchObservedRunningTime="2025-11-29 00:04:34.741028787 +0000 UTC m=+287.902922029" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.743364 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.757073 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.762413 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/85c42636-6a45-439f-9a35-596ae0b40d3d-audit-dir\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.762521 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qswjj\" (UniqueName: \"kubernetes.io/projected/85c42636-6a45-439f-9a35-596ae0b40d3d-kube-api-access-qswjj\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.762550 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-user-template-error\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.762595 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.762638 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.762661 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/85c42636-6a45-439f-9a35-596ae0b40d3d-audit-policies\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.762716 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.762776 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.762804 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-session\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.762887 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-user-template-login\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.762947 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-router-certs\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.763027 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.763073 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-service-ca\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.763100 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.818167 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.863929 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.863979 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-service-ca\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864004 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864047 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/85c42636-6a45-439f-9a35-596ae0b40d3d-audit-dir\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864093 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qswjj\" (UniqueName: \"kubernetes.io/projected/85c42636-6a45-439f-9a35-596ae0b40d3d-kube-api-access-qswjj\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864116 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-user-template-error\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864148 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864175 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864196 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/85c42636-6a45-439f-9a35-596ae0b40d3d-audit-policies\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864221 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864250 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864271 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-session\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864292 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-user-template-login\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864310 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-router-certs\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.864638 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/85c42636-6a45-439f-9a35-596ae0b40d3d-audit-dir\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.865847 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.865900 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-service-ca\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.865971 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/85c42636-6a45-439f-9a35-596ae0b40d3d-audit-policies\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.866388 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.870628 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-session\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.872343 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-user-template-login\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.872615 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.873142 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.873500 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-router-certs\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.874685 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.874881 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-user-template-error\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.877285 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/85c42636-6a45-439f-9a35-596ae0b40d3d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.878166 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.884243 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qswjj\" (UniqueName: \"kubernetes.io/projected/85c42636-6a45-439f-9a35-596ae0b40d3d-kube-api-access-qswjj\") pod \"oauth-openshift-57569d6b9d-stjx8\" (UID: \"85c42636-6a45-439f-9a35-596ae0b40d3d\") " pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.891621 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.894372 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 29 00:04:34 crc kubenswrapper[4931]: I1129 00:04:34.933423 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.007966 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.010793 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.013583 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.031671 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.074309 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.157228 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.188936 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.219267 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.220045 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="199fba37-4269-4fb2-9249-b384fc2f2905" path="/var/lib/kubelet/pods/199fba37-4269-4fb2-9249-b384fc2f2905/volumes" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.259062 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.297704 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.335540 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.385568 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.466566 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.558920 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.575703 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57569d6b9d-stjx8"] Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.638927 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.661609 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.882236 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.919294 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 29 00:04:35 crc kubenswrapper[4931]: I1129 00:04:35.954114 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.000173 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.086959 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.182633 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.214260 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.228888 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.242344 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.267860 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.472779 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.570023 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.716865 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.725974 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.731297 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 29 00:04:36 crc kubenswrapper[4931]: I1129 00:04:36.964966 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.063533 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.207516 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.367720 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.397865 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.412527 4931 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.412839 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.419859 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.437036 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.440425 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.465441 4931 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.519513 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.583053 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.694819 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57569d6b9d-stjx8"] Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.712979 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.734280 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.751396 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.818335 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" event={"ID":"85c42636-6a45-439f-9a35-596ae0b40d3d","Type":"ContainerStarted","Data":"9df4b881d1feca060a1ea00edd21e6073502f281aa5b5c2d1f313e86723d303d"} Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.875249 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 29 00:04:37 crc kubenswrapper[4931]: I1129 00:04:37.938497 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.050426 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.079084 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.221726 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.230732 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.241258 4931 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.283719 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.375582 4931 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.375856 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://31e9ae5daa0e45e16ccea59f5bd0d50977658114846a3705f80646ac07a1c777" gracePeriod=5 Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.408460 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.421759 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.468151 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.518438 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.631647 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.769229 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.823866 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" event={"ID":"85c42636-6a45-439f-9a35-596ae0b40d3d","Type":"ContainerStarted","Data":"ba1cf9ceda06d4038667ccd91a546a5e94adba93ef12cf895fab3113c3262f79"} Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.824386 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.839952 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.854351 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-57569d6b9d-stjx8" podStartSLOduration=63.854337651 podStartE2EDuration="1m3.854337651s" podCreationTimestamp="2025-11-29 00:03:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:04:38.853920828 +0000 UTC m=+292.015814080" watchObservedRunningTime="2025-11-29 00:04:38.854337651 +0000 UTC m=+292.016230883" Nov 29 00:04:38 crc kubenswrapper[4931]: I1129 00:04:38.922176 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.137013 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.184182 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.224265 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.290297 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.292838 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.377008 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.503184 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.561857 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.724283 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.728033 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.812881 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.817585 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 29 00:04:39 crc kubenswrapper[4931]: I1129 00:04:39.963127 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 29 00:04:40 crc kubenswrapper[4931]: I1129 00:04:40.007316 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 29 00:04:40 crc kubenswrapper[4931]: I1129 00:04:40.064504 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 29 00:04:40 crc kubenswrapper[4931]: I1129 00:04:40.248690 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 29 00:04:40 crc kubenswrapper[4931]: I1129 00:04:40.476773 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 29 00:04:40 crc kubenswrapper[4931]: I1129 00:04:40.516880 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 29 00:04:40 crc kubenswrapper[4931]: I1129 00:04:40.539186 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 29 00:04:40 crc kubenswrapper[4931]: I1129 00:04:40.593639 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 29 00:04:40 crc kubenswrapper[4931]: I1129 00:04:40.633446 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 29 00:04:40 crc kubenswrapper[4931]: I1129 00:04:40.777512 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 29 00:04:41 crc kubenswrapper[4931]: I1129 00:04:41.022689 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 29 00:04:41 crc kubenswrapper[4931]: I1129 00:04:41.135494 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 29 00:04:41 crc kubenswrapper[4931]: I1129 00:04:41.197642 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 29 00:04:41 crc kubenswrapper[4931]: I1129 00:04:41.200900 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 29 00:04:41 crc kubenswrapper[4931]: I1129 00:04:41.394726 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 29 00:04:41 crc kubenswrapper[4931]: I1129 00:04:41.545632 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 29 00:04:41 crc kubenswrapper[4931]: I1129 00:04:41.861066 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 29 00:04:42 crc kubenswrapper[4931]: I1129 00:04:42.034343 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 29 00:04:42 crc kubenswrapper[4931]: I1129 00:04:42.151560 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.856894 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.856968 4931 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="31e9ae5daa0e45e16ccea59f5bd0d50977658114846a3705f80646ac07a1c777" exitCode=137 Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.968494 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.968574 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.995727 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.995770 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.995853 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.995898 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.995921 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.995966 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.996050 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.996112 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.996111 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.996192 4931 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.996208 4931 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:43 crc kubenswrapper[4931]: I1129 00:04:43.996220 4931 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:44 crc kubenswrapper[4931]: I1129 00:04:44.007263 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:04:44 crc kubenswrapper[4931]: I1129 00:04:44.097271 4931 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:44 crc kubenswrapper[4931]: I1129 00:04:44.097316 4931 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:44 crc kubenswrapper[4931]: I1129 00:04:44.868331 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 29 00:04:44 crc kubenswrapper[4931]: I1129 00:04:44.868627 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 29 00:04:44 crc kubenswrapper[4931]: I1129 00:04:44.868744 4931 scope.go:117] "RemoveContainer" containerID="31e9ae5daa0e45e16ccea59f5bd0d50977658114846a3705f80646ac07a1c777" Nov 29 00:04:45 crc kubenswrapper[4931]: I1129 00:04:45.219495 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 29 00:04:46 crc kubenswrapper[4931]: I1129 00:04:46.996405 4931 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.056012 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hzhd2"] Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.057168 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hzhd2" podUID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" containerName="registry-server" containerID="cri-o://51bd0d1536062b0b9e4985ed7f272d8a71ee9ad94693d67ef731463be9e3532a" gracePeriod=30 Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.066885 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r5h8c"] Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.067573 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r5h8c" podUID="39557253-7dac-4846-a945-44b03824ec08" containerName="registry-server" containerID="cri-o://3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573" gracePeriod=30 Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.074078 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4zs2z"] Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.074345 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" podUID="62bb1b45-d5f5-4a0d-8c8d-212acc620a3a" containerName="marketplace-operator" containerID="cri-o://f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876" gracePeriod=30 Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.093331 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk72v"] Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.093612 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bk72v" podUID="97c7b21f-8584-482c-b18b-074549e1a120" containerName="registry-server" containerID="cri-o://f84d5d1e015f1a40c54ba5b98bb105100740fe21f39c48211300dba5df23d988" gracePeriod=30 Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.104881 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q9c9x"] Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.105237 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q9c9x" podUID="2e2b7be6-d4e7-4e92-b419-bda83df88970" containerName="registry-server" containerID="cri-o://34168fbe1775d3925b41507a0d49d70ae8af2f7997262babf242e0c6d248bcd3" gracePeriod=30 Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.109053 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m8pmr"] Nov 29 00:04:50 crc kubenswrapper[4931]: E1129 00:04:50.109302 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.109319 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.109448 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.119039 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m8pmr"] Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.119150 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.273640 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8w7k\" (UniqueName: \"kubernetes.io/projected/db3078a6-11fa-4bda-8b8e-d3ac07d1db11-kube-api-access-f8w7k\") pod \"marketplace-operator-79b997595-m8pmr\" (UID: \"db3078a6-11fa-4bda-8b8e-d3ac07d1db11\") " pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.273743 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db3078a6-11fa-4bda-8b8e-d3ac07d1db11-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m8pmr\" (UID: \"db3078a6-11fa-4bda-8b8e-d3ac07d1db11\") " pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.273890 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db3078a6-11fa-4bda-8b8e-d3ac07d1db11-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m8pmr\" (UID: \"db3078a6-11fa-4bda-8b8e-d3ac07d1db11\") " pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.374943 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8w7k\" (UniqueName: \"kubernetes.io/projected/db3078a6-11fa-4bda-8b8e-d3ac07d1db11-kube-api-access-f8w7k\") pod \"marketplace-operator-79b997595-m8pmr\" (UID: \"db3078a6-11fa-4bda-8b8e-d3ac07d1db11\") " pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.374998 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db3078a6-11fa-4bda-8b8e-d3ac07d1db11-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m8pmr\" (UID: \"db3078a6-11fa-4bda-8b8e-d3ac07d1db11\") " pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.375052 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db3078a6-11fa-4bda-8b8e-d3ac07d1db11-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m8pmr\" (UID: \"db3078a6-11fa-4bda-8b8e-d3ac07d1db11\") " pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.376375 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db3078a6-11fa-4bda-8b8e-d3ac07d1db11-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m8pmr\" (UID: \"db3078a6-11fa-4bda-8b8e-d3ac07d1db11\") " pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.388054 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db3078a6-11fa-4bda-8b8e-d3ac07d1db11-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m8pmr\" (UID: \"db3078a6-11fa-4bda-8b8e-d3ac07d1db11\") " pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.401683 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8w7k\" (UniqueName: \"kubernetes.io/projected/db3078a6-11fa-4bda-8b8e-d3ac07d1db11-kube-api-access-f8w7k\") pod \"marketplace-operator-79b997595-m8pmr\" (UID: \"db3078a6-11fa-4bda-8b8e-d3ac07d1db11\") " pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.540743 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.839300 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m8pmr"] Nov 29 00:04:50 crc kubenswrapper[4931]: W1129 00:04:50.846887 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb3078a6_11fa_4bda_8b8e_d3ac07d1db11.slice/crio-a6c66883d2457f48dd4609e52ef25a0419c2cdfb3957fa3da9eddde45b43a34a WatchSource:0}: Error finding container a6c66883d2457f48dd4609e52ef25a0419c2cdfb3957fa3da9eddde45b43a34a: Status 404 returned error can't find the container with id a6c66883d2457f48dd4609e52ef25a0419c2cdfb3957fa3da9eddde45b43a34a Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.905769 4931 generic.go:334] "Generic (PLEG): container finished" podID="97c7b21f-8584-482c-b18b-074549e1a120" containerID="f84d5d1e015f1a40c54ba5b98bb105100740fe21f39c48211300dba5df23d988" exitCode=0 Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.905866 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk72v" event={"ID":"97c7b21f-8584-482c-b18b-074549e1a120","Type":"ContainerDied","Data":"f84d5d1e015f1a40c54ba5b98bb105100740fe21f39c48211300dba5df23d988"} Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.907159 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" event={"ID":"db3078a6-11fa-4bda-8b8e-d3ac07d1db11","Type":"ContainerStarted","Data":"a6c66883d2457f48dd4609e52ef25a0419c2cdfb3957fa3da9eddde45b43a34a"} Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.910297 4931 generic.go:334] "Generic (PLEG): container finished" podID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" containerID="51bd0d1536062b0b9e4985ed7f272d8a71ee9ad94693d67ef731463be9e3532a" exitCode=0 Nov 29 00:04:50 crc kubenswrapper[4931]: I1129 00:04:50.910343 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzhd2" event={"ID":"7c5af5e4-0d88-443e-8f17-93ce08cb4bab","Type":"ContainerDied","Data":"51bd0d1536062b0b9e4985ed7f272d8a71ee9ad94693d67ef731463be9e3532a"} Nov 29 00:04:51 crc kubenswrapper[4931]: E1129 00:04:51.038693 4931 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62bb1b45_d5f5_4a0d_8c8d_212acc620a3a.slice/crio-conmon-f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876.scope\": RecentStats: unable to find data in memory cache]" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.779615 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.815191 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fq6z\" (UniqueName: \"kubernetes.io/projected/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-kube-api-access-6fq6z\") pod \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\" (UID: \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\") " Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.815248 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-utilities\") pod \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\" (UID: \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\") " Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.815281 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-catalog-content\") pod \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\" (UID: \"7c5af5e4-0d88-443e-8f17-93ce08cb4bab\") " Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.820488 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-utilities" (OuterVolumeSpecName: "utilities") pod "7c5af5e4-0d88-443e-8f17-93ce08cb4bab" (UID: "7c5af5e4-0d88-443e-8f17-93ce08cb4bab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.822782 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-kube-api-access-6fq6z" (OuterVolumeSpecName: "kube-api-access-6fq6z") pod "7c5af5e4-0d88-443e-8f17-93ce08cb4bab" (UID: "7c5af5e4-0d88-443e-8f17-93ce08cb4bab"). InnerVolumeSpecName "kube-api-access-6fq6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.855839 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.862997 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.866127 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.876014 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c5af5e4-0d88-443e-8f17-93ce08cb4bab" (UID: "7c5af5e4-0d88-443e-8f17-93ce08cb4bab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.915924 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f65v\" (UniqueName: \"kubernetes.io/projected/39557253-7dac-4846-a945-44b03824ec08-kube-api-access-2f65v\") pod \"39557253-7dac-4846-a945-44b03824ec08\" (UID: \"39557253-7dac-4846-a945-44b03824ec08\") " Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.915963 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-marketplace-operator-metrics\") pod \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\" (UID: \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\") " Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.915995 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqf6h\" (UniqueName: \"kubernetes.io/projected/97c7b21f-8584-482c-b18b-074549e1a120-kube-api-access-pqf6h\") pod \"97c7b21f-8584-482c-b18b-074549e1a120\" (UID: \"97c7b21f-8584-482c-b18b-074549e1a120\") " Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.916017 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zpqp\" (UniqueName: \"kubernetes.io/projected/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-kube-api-access-2zpqp\") pod \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\" (UID: \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\") " Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.916041 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c7b21f-8584-482c-b18b-074549e1a120-catalog-content\") pod \"97c7b21f-8584-482c-b18b-074549e1a120\" (UID: \"97c7b21f-8584-482c-b18b-074549e1a120\") " Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.916079 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c7b21f-8584-482c-b18b-074549e1a120-utilities\") pod \"97c7b21f-8584-482c-b18b-074549e1a120\" (UID: \"97c7b21f-8584-482c-b18b-074549e1a120\") " Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.916100 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-marketplace-trusted-ca\") pod \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\" (UID: \"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a\") " Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.916119 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39557253-7dac-4846-a945-44b03824ec08-utilities\") pod \"39557253-7dac-4846-a945-44b03824ec08\" (UID: \"39557253-7dac-4846-a945-44b03824ec08\") " Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.916168 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39557253-7dac-4846-a945-44b03824ec08-catalog-content\") pod \"39557253-7dac-4846-a945-44b03824ec08\" (UID: \"39557253-7dac-4846-a945-44b03824ec08\") " Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.916365 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.916376 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.916385 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fq6z\" (UniqueName: \"kubernetes.io/projected/7c5af5e4-0d88-443e-8f17-93ce08cb4bab-kube-api-access-6fq6z\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.918495 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "62bb1b45-d5f5-4a0d-8c8d-212acc620a3a" (UID: "62bb1b45-d5f5-4a0d-8c8d-212acc620a3a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.919585 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39557253-7dac-4846-a945-44b03824ec08-utilities" (OuterVolumeSpecName: "utilities") pod "39557253-7dac-4846-a945-44b03824ec08" (UID: "39557253-7dac-4846-a945-44b03824ec08"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.919684 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97c7b21f-8584-482c-b18b-074549e1a120-kube-api-access-pqf6h" (OuterVolumeSpecName: "kube-api-access-pqf6h") pod "97c7b21f-8584-482c-b18b-074549e1a120" (UID: "97c7b21f-8584-482c-b18b-074549e1a120"). InnerVolumeSpecName "kube-api-access-pqf6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.919946 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97c7b21f-8584-482c-b18b-074549e1a120-utilities" (OuterVolumeSpecName: "utilities") pod "97c7b21f-8584-482c-b18b-074549e1a120" (UID: "97c7b21f-8584-482c-b18b-074549e1a120"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.921609 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "62bb1b45-d5f5-4a0d-8c8d-212acc620a3a" (UID: "62bb1b45-d5f5-4a0d-8c8d-212acc620a3a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.922573 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39557253-7dac-4846-a945-44b03824ec08-kube-api-access-2f65v" (OuterVolumeSpecName: "kube-api-access-2f65v") pod "39557253-7dac-4846-a945-44b03824ec08" (UID: "39557253-7dac-4846-a945-44b03824ec08"). InnerVolumeSpecName "kube-api-access-2f65v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.924835 4931 generic.go:334] "Generic (PLEG): container finished" podID="39557253-7dac-4846-a945-44b03824ec08" containerID="3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573" exitCode=0 Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.924996 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r5h8c" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.926175 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5h8c" event={"ID":"39557253-7dac-4846-a945-44b03824ec08","Type":"ContainerDied","Data":"3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573"} Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.926217 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r5h8c" event={"ID":"39557253-7dac-4846-a945-44b03824ec08","Type":"ContainerDied","Data":"fcb95e718e5ac685f08eca6b7746cc1ff3e8632687e106739395060f11bb1ca6"} Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.926239 4931 scope.go:117] "RemoveContainer" containerID="3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.935687 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-kube-api-access-2zpqp" (OuterVolumeSpecName: "kube-api-access-2zpqp") pod "62bb1b45-d5f5-4a0d-8c8d-212acc620a3a" (UID: "62bb1b45-d5f5-4a0d-8c8d-212acc620a3a"). InnerVolumeSpecName "kube-api-access-2zpqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.937066 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bk72v" event={"ID":"97c7b21f-8584-482c-b18b-074549e1a120","Type":"ContainerDied","Data":"304d4632cdbe4bebd8359321539d47d8f14a710d1516d64137a6e7803fbd5ca6"} Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.937234 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bk72v" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.949245 4931 generic.go:334] "Generic (PLEG): container finished" podID="2e2b7be6-d4e7-4e92-b419-bda83df88970" containerID="34168fbe1775d3925b41507a0d49d70ae8af2f7997262babf242e0c6d248bcd3" exitCode=0 Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.949348 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9c9x" event={"ID":"2e2b7be6-d4e7-4e92-b419-bda83df88970","Type":"ContainerDied","Data":"34168fbe1775d3925b41507a0d49d70ae8af2f7997262babf242e0c6d248bcd3"} Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.950780 4931 scope.go:117] "RemoveContainer" containerID="37f9cb08e81eb8bf7642bfd1d7d203d9ca104b67b40893273731704e39148fdd" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.965202 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" event={"ID":"db3078a6-11fa-4bda-8b8e-d3ac07d1db11","Type":"ContainerStarted","Data":"ed08da5863a8cd9ee3b9641e48a63ecc92876b3a76ea81b0027822eeb23e87fb"} Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.966026 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.973753 4931 generic.go:334] "Generic (PLEG): container finished" podID="62bb1b45-d5f5-4a0d-8c8d-212acc620a3a" containerID="f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876" exitCode=0 Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.973860 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" event={"ID":"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a","Type":"ContainerDied","Data":"f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876"} Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.973887 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" event={"ID":"62bb1b45-d5f5-4a0d-8c8d-212acc620a3a","Type":"ContainerDied","Data":"7e61c3011cf88f7cab6824f1504f670559e1a17e78910d7175bb15249f5c1e03"} Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.973953 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4zs2z" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.972271 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97c7b21f-8584-482c-b18b-074549e1a120-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97c7b21f-8584-482c-b18b-074549e1a120" (UID: "97c7b21f-8584-482c-b18b-074549e1a120"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.983100 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.987553 4931 scope.go:117] "RemoveContainer" containerID="9d6789c2fc8f197dbc1f43572bd0f93b8fa624e96c1f0fd244cbaee3aaf9af63" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.988670 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-m8pmr" podStartSLOduration=1.988649978 podStartE2EDuration="1.988649978s" podCreationTimestamp="2025-11-29 00:04:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:04:51.986904715 +0000 UTC m=+305.148797957" watchObservedRunningTime="2025-11-29 00:04:51.988649978 +0000 UTC m=+305.150543220" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.989281 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzhd2" event={"ID":"7c5af5e4-0d88-443e-8f17-93ce08cb4bab","Type":"ContainerDied","Data":"1b05cbbae8781a790627c92e494b3b5b0edac37d067f55a3ac9fd12aee1cf87b"} Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.989365 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hzhd2" Nov 29 00:04:51 crc kubenswrapper[4931]: I1129 00:04:51.989826 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39557253-7dac-4846-a945-44b03824ec08-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39557253-7dac-4846-a945-44b03824ec08" (UID: "39557253-7dac-4846-a945-44b03824ec08"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.014149 4931 scope.go:117] "RemoveContainer" containerID="3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573" Nov 29 00:04:52 crc kubenswrapper[4931]: E1129 00:04:52.015052 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573\": container with ID starting with 3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573 not found: ID does not exist" containerID="3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.015083 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573"} err="failed to get container status \"3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573\": rpc error: code = NotFound desc = could not find container \"3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573\": container with ID starting with 3056a38f995230ef304ecb6993fddfc8877b62dce32c2e548504a02a38db7573 not found: ID does not exist" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.015105 4931 scope.go:117] "RemoveContainer" containerID="37f9cb08e81eb8bf7642bfd1d7d203d9ca104b67b40893273731704e39148fdd" Nov 29 00:04:52 crc kubenswrapper[4931]: E1129 00:04:52.015374 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37f9cb08e81eb8bf7642bfd1d7d203d9ca104b67b40893273731704e39148fdd\": container with ID starting with 37f9cb08e81eb8bf7642bfd1d7d203d9ca104b67b40893273731704e39148fdd not found: ID does not exist" containerID="37f9cb08e81eb8bf7642bfd1d7d203d9ca104b67b40893273731704e39148fdd" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.015392 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f9cb08e81eb8bf7642bfd1d7d203d9ca104b67b40893273731704e39148fdd"} err="failed to get container status \"37f9cb08e81eb8bf7642bfd1d7d203d9ca104b67b40893273731704e39148fdd\": rpc error: code = NotFound desc = could not find container \"37f9cb08e81eb8bf7642bfd1d7d203d9ca104b67b40893273731704e39148fdd\": container with ID starting with 37f9cb08e81eb8bf7642bfd1d7d203d9ca104b67b40893273731704e39148fdd not found: ID does not exist" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.015420 4931 scope.go:117] "RemoveContainer" containerID="9d6789c2fc8f197dbc1f43572bd0f93b8fa624e96c1f0fd244cbaee3aaf9af63" Nov 29 00:04:52 crc kubenswrapper[4931]: E1129 00:04:52.015670 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d6789c2fc8f197dbc1f43572bd0f93b8fa624e96c1f0fd244cbaee3aaf9af63\": container with ID starting with 9d6789c2fc8f197dbc1f43572bd0f93b8fa624e96c1f0fd244cbaee3aaf9af63 not found: ID does not exist" containerID="9d6789c2fc8f197dbc1f43572bd0f93b8fa624e96c1f0fd244cbaee3aaf9af63" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.015735 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d6789c2fc8f197dbc1f43572bd0f93b8fa624e96c1f0fd244cbaee3aaf9af63"} err="failed to get container status \"9d6789c2fc8f197dbc1f43572bd0f93b8fa624e96c1f0fd244cbaee3aaf9af63\": rpc error: code = NotFound desc = could not find container \"9d6789c2fc8f197dbc1f43572bd0f93b8fa624e96c1f0fd244cbaee3aaf9af63\": container with ID starting with 9d6789c2fc8f197dbc1f43572bd0f93b8fa624e96c1f0fd244cbaee3aaf9af63 not found: ID does not exist" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.015764 4931 scope.go:117] "RemoveContainer" containerID="f84d5d1e015f1a40c54ba5b98bb105100740fe21f39c48211300dba5df23d988" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.017152 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c7b21f-8584-482c-b18b-074549e1a120-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.017176 4931 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.017185 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39557253-7dac-4846-a945-44b03824ec08-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.017195 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39557253-7dac-4846-a945-44b03824ec08-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.017204 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f65v\" (UniqueName: \"kubernetes.io/projected/39557253-7dac-4846-a945-44b03824ec08-kube-api-access-2f65v\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.017212 4931 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.017222 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqf6h\" (UniqueName: \"kubernetes.io/projected/97c7b21f-8584-482c-b18b-074549e1a120-kube-api-access-pqf6h\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.017230 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zpqp\" (UniqueName: \"kubernetes.io/projected/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a-kube-api-access-2zpqp\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.017239 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c7b21f-8584-482c-b18b-074549e1a120-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.029075 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hzhd2"] Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.035588 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hzhd2"] Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.040714 4931 scope.go:117] "RemoveContainer" containerID="a4a6bb2a3b6aad78d54d32e58766ff0a7e8de757329e3fd25bb455dd5f6b4d15" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.040869 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4zs2z"] Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.045170 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4zs2z"] Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.052136 4931 scope.go:117] "RemoveContainer" containerID="5a4d7fb7165df19636715d713472577e37de60837cc92eadbe2112b024524b25" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.062269 4931 scope.go:117] "RemoveContainer" containerID="f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.073528 4931 scope.go:117] "RemoveContainer" containerID="f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876" Nov 29 00:04:52 crc kubenswrapper[4931]: E1129 00:04:52.074048 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876\": container with ID starting with f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876 not found: ID does not exist" containerID="f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.074101 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876"} err="failed to get container status \"f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876\": rpc error: code = NotFound desc = could not find container \"f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876\": container with ID starting with f4774f444587108b661daa992eeb56896747690ba51fbd1307bdeafcf314a876 not found: ID does not exist" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.074127 4931 scope.go:117] "RemoveContainer" containerID="51bd0d1536062b0b9e4985ed7f272d8a71ee9ad94693d67ef731463be9e3532a" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.092202 4931 scope.go:117] "RemoveContainer" containerID="d996a4e5ae36ffe8bad8ef0312f727e651e25d79e1c6e5cb0a652084b08c3a65" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.104459 4931 scope.go:117] "RemoveContainer" containerID="b5d2161d19f8b120d4d8c56a2c734b8b4bd1cfb7df591debf08c953a1378fd44" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.271793 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r5h8c"] Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.283541 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r5h8c"] Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.289750 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk72v"] Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.295182 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bk72v"] Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.892629 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.927261 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e2b7be6-d4e7-4e92-b419-bda83df88970-utilities\") pod \"2e2b7be6-d4e7-4e92-b419-bda83df88970\" (UID: \"2e2b7be6-d4e7-4e92-b419-bda83df88970\") " Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.927337 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e2b7be6-d4e7-4e92-b419-bda83df88970-catalog-content\") pod \"2e2b7be6-d4e7-4e92-b419-bda83df88970\" (UID: \"2e2b7be6-d4e7-4e92-b419-bda83df88970\") " Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.927412 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6f2s\" (UniqueName: \"kubernetes.io/projected/2e2b7be6-d4e7-4e92-b419-bda83df88970-kube-api-access-p6f2s\") pod \"2e2b7be6-d4e7-4e92-b419-bda83df88970\" (UID: \"2e2b7be6-d4e7-4e92-b419-bda83df88970\") " Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.928180 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e2b7be6-d4e7-4e92-b419-bda83df88970-utilities" (OuterVolumeSpecName: "utilities") pod "2e2b7be6-d4e7-4e92-b419-bda83df88970" (UID: "2e2b7be6-d4e7-4e92-b419-bda83df88970"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:04:52 crc kubenswrapper[4931]: I1129 00:04:52.934140 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e2b7be6-d4e7-4e92-b419-bda83df88970-kube-api-access-p6f2s" (OuterVolumeSpecName: "kube-api-access-p6f2s") pod "2e2b7be6-d4e7-4e92-b419-bda83df88970" (UID: "2e2b7be6-d4e7-4e92-b419-bda83df88970"). InnerVolumeSpecName "kube-api-access-p6f2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.000492 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9c9x" event={"ID":"2e2b7be6-d4e7-4e92-b419-bda83df88970","Type":"ContainerDied","Data":"cec7b0fa99c7095c785da0e358bc1d27ad2dfa3c30cd2ff81167ba8543e53ce3"} Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.000537 4931 scope.go:117] "RemoveContainer" containerID="34168fbe1775d3925b41507a0d49d70ae8af2f7997262babf242e0c6d248bcd3" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.000546 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9c9x" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.015305 4931 scope.go:117] "RemoveContainer" containerID="88a7b08d2e0dc9d9c1ae3969d79d20fbdf95bfa0080d33085ce90d4db9a828f1" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.028846 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e2b7be6-d4e7-4e92-b419-bda83df88970-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.029034 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6f2s\" (UniqueName: \"kubernetes.io/projected/2e2b7be6-d4e7-4e92-b419-bda83df88970-kube-api-access-p6f2s\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.030880 4931 scope.go:117] "RemoveContainer" containerID="39b3341548aea9fbb9cc2e6501a0b834580047b5ec4d94faf1f3103146718472" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.042829 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e2b7be6-d4e7-4e92-b419-bda83df88970-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e2b7be6-d4e7-4e92-b419-bda83df88970" (UID: "2e2b7be6-d4e7-4e92-b419-bda83df88970"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.130529 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e2b7be6-d4e7-4e92-b419-bda83df88970-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.225321 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39557253-7dac-4846-a945-44b03824ec08" path="/var/lib/kubelet/pods/39557253-7dac-4846-a945-44b03824ec08/volumes" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.226972 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62bb1b45-d5f5-4a0d-8c8d-212acc620a3a" path="/var/lib/kubelet/pods/62bb1b45-d5f5-4a0d-8c8d-212acc620a3a/volumes" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.228107 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" path="/var/lib/kubelet/pods/7c5af5e4-0d88-443e-8f17-93ce08cb4bab/volumes" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.230635 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97c7b21f-8584-482c-b18b-074549e1a120" path="/var/lib/kubelet/pods/97c7b21f-8584-482c-b18b-074549e1a120/volumes" Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.329027 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q9c9x"] Nov 29 00:04:53 crc kubenswrapper[4931]: I1129 00:04:53.337289 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q9c9x"] Nov 29 00:04:55 crc kubenswrapper[4931]: I1129 00:04:55.223095 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e2b7be6-d4e7-4e92-b419-bda83df88970" path="/var/lib/kubelet/pods/2e2b7be6-d4e7-4e92-b419-bda83df88970/volumes" Nov 29 00:04:56 crc kubenswrapper[4931]: I1129 00:04:56.305827 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 29 00:04:56 crc kubenswrapper[4931]: I1129 00:04:56.430345 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.190607 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ffwsb"] Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.191210 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" podUID="13b49f2e-d3eb-483c-a910-202c955bfda9" containerName="controller-manager" containerID="cri-o://73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66" gracePeriod=30 Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.289419 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf"] Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.289621 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" podUID="0be48302-a154-4099-b4d7-3e4781964c7d" containerName="route-controller-manager" containerID="cri-o://5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526" gracePeriod=30 Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.686509 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.691588 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.715096 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbgrn\" (UniqueName: \"kubernetes.io/projected/13b49f2e-d3eb-483c-a910-202c955bfda9-kube-api-access-tbgrn\") pod \"13b49f2e-d3eb-483c-a910-202c955bfda9\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.715165 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-config\") pod \"13b49f2e-d3eb-483c-a910-202c955bfda9\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.715200 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0be48302-a154-4099-b4d7-3e4781964c7d-serving-cert\") pod \"0be48302-a154-4099-b4d7-3e4781964c7d\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.715262 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be48302-a154-4099-b4d7-3e4781964c7d-config\") pod \"0be48302-a154-4099-b4d7-3e4781964c7d\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.715282 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-client-ca\") pod \"13b49f2e-d3eb-483c-a910-202c955bfda9\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.715306 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13b49f2e-d3eb-483c-a910-202c955bfda9-serving-cert\") pod \"13b49f2e-d3eb-483c-a910-202c955bfda9\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.715334 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2vmm\" (UniqueName: \"kubernetes.io/projected/0be48302-a154-4099-b4d7-3e4781964c7d-kube-api-access-v2vmm\") pod \"0be48302-a154-4099-b4d7-3e4781964c7d\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.715363 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0be48302-a154-4099-b4d7-3e4781964c7d-client-ca\") pod \"0be48302-a154-4099-b4d7-3e4781964c7d\" (UID: \"0be48302-a154-4099-b4d7-3e4781964c7d\") " Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.715387 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-proxy-ca-bundles\") pod \"13b49f2e-d3eb-483c-a910-202c955bfda9\" (UID: \"13b49f2e-d3eb-483c-a910-202c955bfda9\") " Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.716402 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "13b49f2e-d3eb-483c-a910-202c955bfda9" (UID: "13b49f2e-d3eb-483c-a910-202c955bfda9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.717499 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-client-ca" (OuterVolumeSpecName: "client-ca") pod "13b49f2e-d3eb-483c-a910-202c955bfda9" (UID: "13b49f2e-d3eb-483c-a910-202c955bfda9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.717919 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-config" (OuterVolumeSpecName: "config") pod "13b49f2e-d3eb-483c-a910-202c955bfda9" (UID: "13b49f2e-d3eb-483c-a910-202c955bfda9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.719347 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0be48302-a154-4099-b4d7-3e4781964c7d-config" (OuterVolumeSpecName: "config") pod "0be48302-a154-4099-b4d7-3e4781964c7d" (UID: "0be48302-a154-4099-b4d7-3e4781964c7d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.719686 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0be48302-a154-4099-b4d7-3e4781964c7d-client-ca" (OuterVolumeSpecName: "client-ca") pod "0be48302-a154-4099-b4d7-3e4781964c7d" (UID: "0be48302-a154-4099-b4d7-3e4781964c7d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.729521 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0be48302-a154-4099-b4d7-3e4781964c7d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0be48302-a154-4099-b4d7-3e4781964c7d" (UID: "0be48302-a154-4099-b4d7-3e4781964c7d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.729550 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13b49f2e-d3eb-483c-a910-202c955bfda9-kube-api-access-tbgrn" (OuterVolumeSpecName: "kube-api-access-tbgrn") pod "13b49f2e-d3eb-483c-a910-202c955bfda9" (UID: "13b49f2e-d3eb-483c-a910-202c955bfda9"). InnerVolumeSpecName "kube-api-access-tbgrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.729530 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13b49f2e-d3eb-483c-a910-202c955bfda9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "13b49f2e-d3eb-483c-a910-202c955bfda9" (UID: "13b49f2e-d3eb-483c-a910-202c955bfda9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.729663 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0be48302-a154-4099-b4d7-3e4781964c7d-kube-api-access-v2vmm" (OuterVolumeSpecName: "kube-api-access-v2vmm") pod "0be48302-a154-4099-b4d7-3e4781964c7d" (UID: "0be48302-a154-4099-b4d7-3e4781964c7d"). InnerVolumeSpecName "kube-api-access-v2vmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.817140 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.817184 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0be48302-a154-4099-b4d7-3e4781964c7d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.817197 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0be48302-a154-4099-b4d7-3e4781964c7d-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.817208 4931 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.817220 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13b49f2e-d3eb-483c-a910-202c955bfda9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.817232 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2vmm\" (UniqueName: \"kubernetes.io/projected/0be48302-a154-4099-b4d7-3e4781964c7d-kube-api-access-v2vmm\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.817248 4931 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0be48302-a154-4099-b4d7-3e4781964c7d-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.817259 4931 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/13b49f2e-d3eb-483c-a910-202c955bfda9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:06 crc kubenswrapper[4931]: I1129 00:05:06.817272 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbgrn\" (UniqueName: \"kubernetes.io/projected/13b49f2e-d3eb-483c-a910-202c955bfda9-kube-api-access-tbgrn\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.112441 4931 generic.go:334] "Generic (PLEG): container finished" podID="0be48302-a154-4099-b4d7-3e4781964c7d" containerID="5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526" exitCode=0 Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.112512 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.112512 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" event={"ID":"0be48302-a154-4099-b4d7-3e4781964c7d","Type":"ContainerDied","Data":"5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526"} Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.112691 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf" event={"ID":"0be48302-a154-4099-b4d7-3e4781964c7d","Type":"ContainerDied","Data":"578a78f40900e9f24d3acfbdc6f00519b470506d18c38261eeb0867104620453"} Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.112748 4931 scope.go:117] "RemoveContainer" containerID="5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.114578 4931 generic.go:334] "Generic (PLEG): container finished" podID="13b49f2e-d3eb-483c-a910-202c955bfda9" containerID="73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66" exitCode=0 Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.114613 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.114629 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" event={"ID":"13b49f2e-d3eb-483c-a910-202c955bfda9","Type":"ContainerDied","Data":"73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66"} Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.114684 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ffwsb" event={"ID":"13b49f2e-d3eb-483c-a910-202c955bfda9","Type":"ContainerDied","Data":"8c3ef93cdb8373c42212ea35ddaae366733ffdc5fafa935c984ce36295945551"} Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.133721 4931 scope.go:117] "RemoveContainer" containerID="5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.134223 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526\": container with ID starting with 5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526 not found: ID does not exist" containerID="5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.134256 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526"} err="failed to get container status \"5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526\": rpc error: code = NotFound desc = could not find container \"5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526\": container with ID starting with 5b70ec95f446695087ab8fbf7536b59d19fe7772fd6650b289c1e2784f928526 not found: ID does not exist" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.134274 4931 scope.go:117] "RemoveContainer" containerID="73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.143669 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ffwsb"] Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.147263 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ffwsb"] Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.159846 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf"] Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.162947 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rqlwf"] Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.163735 4931 scope.go:117] "RemoveContainer" containerID="73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.164197 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66\": container with ID starting with 73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66 not found: ID does not exist" containerID="73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.164238 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66"} err="failed to get container status \"73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66\": rpc error: code = NotFound desc = could not find container \"73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66\": container with ID starting with 73759c8baa76b155677f7f432a204363cdcabdd6434eafc5fa0875b4fb199d66 not found: ID does not exist" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.220988 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0be48302-a154-4099-b4d7-3e4781964c7d" path="/var/lib/kubelet/pods/0be48302-a154-4099-b4d7-3e4781964c7d/volumes" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.222115 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13b49f2e-d3eb-483c-a910-202c955bfda9" path="/var/lib/kubelet/pods/13b49f2e-d3eb-483c-a910-202c955bfda9/volumes" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.481470 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb"] Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.481853 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62bb1b45-d5f5-4a0d-8c8d-212acc620a3a" containerName="marketplace-operator" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.481874 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="62bb1b45-d5f5-4a0d-8c8d-212acc620a3a" containerName="marketplace-operator" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.481890 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0be48302-a154-4099-b4d7-3e4781964c7d" containerName="route-controller-manager" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.481901 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="0be48302-a154-4099-b4d7-3e4781964c7d" containerName="route-controller-manager" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.481914 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39557253-7dac-4846-a945-44b03824ec08" containerName="extract-content" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.481922 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="39557253-7dac-4846-a945-44b03824ec08" containerName="extract-content" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.481937 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" containerName="extract-content" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.481946 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" containerName="extract-content" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.481957 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c7b21f-8584-482c-b18b-074549e1a120" containerName="registry-server" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.481965 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c7b21f-8584-482c-b18b-074549e1a120" containerName="registry-server" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.481977 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13b49f2e-d3eb-483c-a910-202c955bfda9" containerName="controller-manager" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.481986 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="13b49f2e-d3eb-483c-a910-202c955bfda9" containerName="controller-manager" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.481996 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2b7be6-d4e7-4e92-b419-bda83df88970" containerName="extract-content" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482005 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2b7be6-d4e7-4e92-b419-bda83df88970" containerName="extract-content" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.482020 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c7b21f-8584-482c-b18b-074549e1a120" containerName="extract-utilities" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482028 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c7b21f-8584-482c-b18b-074549e1a120" containerName="extract-utilities" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.482040 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c7b21f-8584-482c-b18b-074549e1a120" containerName="extract-content" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482049 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c7b21f-8584-482c-b18b-074549e1a120" containerName="extract-content" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.482426 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" containerName="extract-utilities" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482443 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" containerName="extract-utilities" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.482456 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39557253-7dac-4846-a945-44b03824ec08" containerName="registry-server" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482465 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="39557253-7dac-4846-a945-44b03824ec08" containerName="registry-server" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.482479 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2b7be6-d4e7-4e92-b419-bda83df88970" containerName="extract-utilities" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482488 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2b7be6-d4e7-4e92-b419-bda83df88970" containerName="extract-utilities" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.482499 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" containerName="registry-server" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482507 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" containerName="registry-server" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.482520 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2b7be6-d4e7-4e92-b419-bda83df88970" containerName="registry-server" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482564 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2b7be6-d4e7-4e92-b419-bda83df88970" containerName="registry-server" Nov 29 00:05:07 crc kubenswrapper[4931]: E1129 00:05:07.482579 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39557253-7dac-4846-a945-44b03824ec08" containerName="extract-utilities" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482646 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="39557253-7dac-4846-a945-44b03824ec08" containerName="extract-utilities" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482879 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="13b49f2e-d3eb-483c-a910-202c955bfda9" containerName="controller-manager" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482901 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c5af5e4-0d88-443e-8f17-93ce08cb4bab" containerName="registry-server" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482955 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="62bb1b45-d5f5-4a0d-8c8d-212acc620a3a" containerName="marketplace-operator" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.482967 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="0be48302-a154-4099-b4d7-3e4781964c7d" containerName="route-controller-manager" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.483023 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="97c7b21f-8584-482c-b18b-074549e1a120" containerName="registry-server" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.483104 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="39557253-7dac-4846-a945-44b03824ec08" containerName="registry-server" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.483121 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e2b7be6-d4e7-4e92-b419-bda83df88970" containerName="registry-server" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.483664 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.485945 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-56f9c54474-44w8q"] Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.486357 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.486680 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.486787 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.487530 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.487753 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.488799 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.488987 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.492908 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.493111 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.495940 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.495985 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.496394 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.496403 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.499115 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56f9c54474-44w8q"] Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.504406 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb"] Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.512016 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.550033 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsr7l\" (UniqueName: \"kubernetes.io/projected/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-kube-api-access-tsr7l\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.551934 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-proxy-ca-bundles\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.552096 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-serving-cert\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.552270 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-client-ca\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.552445 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-config\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.653137 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-client-ca\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.653265 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-config\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.654800 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-client-ca\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.657040 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h75xg\" (UniqueName: \"kubernetes.io/projected/1fa096fe-9cff-4481-8e50-6eeb942b2354-kube-api-access-h75xg\") pod \"route-controller-manager-6f5b6b6f74-rt9mb\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.657096 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsr7l\" (UniqueName: \"kubernetes.io/projected/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-kube-api-access-tsr7l\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.657120 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa096fe-9cff-4481-8e50-6eeb942b2354-config\") pod \"route-controller-manager-6f5b6b6f74-rt9mb\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.657210 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-proxy-ca-bundles\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.657242 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-serving-cert\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.657282 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fa096fe-9cff-4481-8e50-6eeb942b2354-serving-cert\") pod \"route-controller-manager-6f5b6b6f74-rt9mb\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.657301 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1fa096fe-9cff-4481-8e50-6eeb942b2354-client-ca\") pod \"route-controller-manager-6f5b6b6f74-rt9mb\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.658945 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-config\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.659233 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-proxy-ca-bundles\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.663300 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-serving-cert\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.688973 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsr7l\" (UniqueName: \"kubernetes.io/projected/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-kube-api-access-tsr7l\") pod \"controller-manager-56f9c54474-44w8q\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.759139 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h75xg\" (UniqueName: \"kubernetes.io/projected/1fa096fe-9cff-4481-8e50-6eeb942b2354-kube-api-access-h75xg\") pod \"route-controller-manager-6f5b6b6f74-rt9mb\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.759939 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa096fe-9cff-4481-8e50-6eeb942b2354-config\") pod \"route-controller-manager-6f5b6b6f74-rt9mb\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.762024 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa096fe-9cff-4481-8e50-6eeb942b2354-config\") pod \"route-controller-manager-6f5b6b6f74-rt9mb\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.762183 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fa096fe-9cff-4481-8e50-6eeb942b2354-serving-cert\") pod \"route-controller-manager-6f5b6b6f74-rt9mb\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.762263 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1fa096fe-9cff-4481-8e50-6eeb942b2354-client-ca\") pod \"route-controller-manager-6f5b6b6f74-rt9mb\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.763507 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1fa096fe-9cff-4481-8e50-6eeb942b2354-client-ca\") pod \"route-controller-manager-6f5b6b6f74-rt9mb\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.767187 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fa096fe-9cff-4481-8e50-6eeb942b2354-serving-cert\") pod \"route-controller-manager-6f5b6b6f74-rt9mb\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.775025 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h75xg\" (UniqueName: \"kubernetes.io/projected/1fa096fe-9cff-4481-8e50-6eeb942b2354-kube-api-access-h75xg\") pod \"route-controller-manager-6f5b6b6f74-rt9mb\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.864461 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:07 crc kubenswrapper[4931]: I1129 00:05:07.871908 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:08 crc kubenswrapper[4931]: I1129 00:05:08.074060 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb"] Nov 29 00:05:08 crc kubenswrapper[4931]: I1129 00:05:08.131720 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" event={"ID":"1fa096fe-9cff-4481-8e50-6eeb942b2354","Type":"ContainerStarted","Data":"8cc4320458669171156861a22cb46c73eba169c24cbf5883f3c4a34ebb49f206"} Nov 29 00:05:08 crc kubenswrapper[4931]: I1129 00:05:08.139413 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56f9c54474-44w8q"] Nov 29 00:05:08 crc kubenswrapper[4931]: I1129 00:05:08.968786 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 29 00:05:09 crc kubenswrapper[4931]: I1129 00:05:09.137515 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" event={"ID":"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6","Type":"ContainerStarted","Data":"f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc"} Nov 29 00:05:09 crc kubenswrapper[4931]: I1129 00:05:09.137602 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" event={"ID":"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6","Type":"ContainerStarted","Data":"7ecabf962d43ad58ded13d9b581178d36c54e2f759f5e2a4f3861ecb1d53dec7"} Nov 29 00:05:09 crc kubenswrapper[4931]: I1129 00:05:09.137970 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:09 crc kubenswrapper[4931]: I1129 00:05:09.140069 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" event={"ID":"1fa096fe-9cff-4481-8e50-6eeb942b2354","Type":"ContainerStarted","Data":"43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5"} Nov 29 00:05:09 crc kubenswrapper[4931]: I1129 00:05:09.140241 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:09 crc kubenswrapper[4931]: I1129 00:05:09.142076 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:09 crc kubenswrapper[4931]: I1129 00:05:09.180664 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" podStartSLOduration=3.180642694 podStartE2EDuration="3.180642694s" podCreationTimestamp="2025-11-29 00:05:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:05:09.157671738 +0000 UTC m=+322.319564990" watchObservedRunningTime="2025-11-29 00:05:09.180642694 +0000 UTC m=+322.342535926" Nov 29 00:05:09 crc kubenswrapper[4931]: I1129 00:05:09.209958 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" podStartSLOduration=3.209940171 podStartE2EDuration="3.209940171s" podCreationTimestamp="2025-11-29 00:05:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:05:09.206067174 +0000 UTC m=+322.367960406" watchObservedRunningTime="2025-11-29 00:05:09.209940171 +0000 UTC m=+322.371833403" Nov 29 00:05:09 crc kubenswrapper[4931]: I1129 00:05:09.423362 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.485723 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zn6ld"] Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.486794 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.488942 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.497062 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zn6ld"] Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.542962 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92e6619-370d-46ea-8925-a35dff600120-catalog-content\") pod \"community-operators-zn6ld\" (UID: \"c92e6619-370d-46ea-8925-a35dff600120\") " pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.543091 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnjhp\" (UniqueName: \"kubernetes.io/projected/c92e6619-370d-46ea-8925-a35dff600120-kube-api-access-hnjhp\") pod \"community-operators-zn6ld\" (UID: \"c92e6619-370d-46ea-8925-a35dff600120\") " pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.543151 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92e6619-370d-46ea-8925-a35dff600120-utilities\") pod \"community-operators-zn6ld\" (UID: \"c92e6619-370d-46ea-8925-a35dff600120\") " pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.643739 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92e6619-370d-46ea-8925-a35dff600120-catalog-content\") pod \"community-operators-zn6ld\" (UID: \"c92e6619-370d-46ea-8925-a35dff600120\") " pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.643791 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnjhp\" (UniqueName: \"kubernetes.io/projected/c92e6619-370d-46ea-8925-a35dff600120-kube-api-access-hnjhp\") pod \"community-operators-zn6ld\" (UID: \"c92e6619-370d-46ea-8925-a35dff600120\") " pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.643835 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92e6619-370d-46ea-8925-a35dff600120-utilities\") pod \"community-operators-zn6ld\" (UID: \"c92e6619-370d-46ea-8925-a35dff600120\") " pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.644332 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92e6619-370d-46ea-8925-a35dff600120-catalog-content\") pod \"community-operators-zn6ld\" (UID: \"c92e6619-370d-46ea-8925-a35dff600120\") " pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.644373 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92e6619-370d-46ea-8925-a35dff600120-utilities\") pod \"community-operators-zn6ld\" (UID: \"c92e6619-370d-46ea-8925-a35dff600120\") " pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.662523 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnjhp\" (UniqueName: \"kubernetes.io/projected/c92e6619-370d-46ea-8925-a35dff600120-kube-api-access-hnjhp\") pod \"community-operators-zn6ld\" (UID: \"c92e6619-370d-46ea-8925-a35dff600120\") " pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.685884 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jwvrh"] Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.687054 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.690199 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.695089 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jwvrh"] Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.744928 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6-utilities\") pod \"redhat-marketplace-jwvrh\" (UID: \"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6\") " pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.744976 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6-catalog-content\") pod \"redhat-marketplace-jwvrh\" (UID: \"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6\") " pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.745011 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg9rd\" (UniqueName: \"kubernetes.io/projected/b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6-kube-api-access-tg9rd\") pod \"redhat-marketplace-jwvrh\" (UID: \"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6\") " pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.844179 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.845697 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6-utilities\") pod \"redhat-marketplace-jwvrh\" (UID: \"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6\") " pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.845737 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6-catalog-content\") pod \"redhat-marketplace-jwvrh\" (UID: \"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6\") " pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.845771 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg9rd\" (UniqueName: \"kubernetes.io/projected/b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6-kube-api-access-tg9rd\") pod \"redhat-marketplace-jwvrh\" (UID: \"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6\") " pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.846246 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6-utilities\") pod \"redhat-marketplace-jwvrh\" (UID: \"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6\") " pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.846246 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6-catalog-content\") pod \"redhat-marketplace-jwvrh\" (UID: \"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6\") " pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:12 crc kubenswrapper[4931]: I1129 00:05:12.862421 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg9rd\" (UniqueName: \"kubernetes.io/projected/b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6-kube-api-access-tg9rd\") pod \"redhat-marketplace-jwvrh\" (UID: \"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6\") " pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:13 crc kubenswrapper[4931]: I1129 00:05:13.002035 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:13 crc kubenswrapper[4931]: I1129 00:05:13.268515 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zn6ld"] Nov 29 00:05:13 crc kubenswrapper[4931]: I1129 00:05:13.399799 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jwvrh"] Nov 29 00:05:13 crc kubenswrapper[4931]: W1129 00:05:13.404963 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3a10cb9_e75c_45e6_a51b_4b6fbeb893c6.slice/crio-4cf59c9455cbddd5f45fb450a94f113fd2c5a9979ac1ab9093021252436f60f2 WatchSource:0}: Error finding container 4cf59c9455cbddd5f45fb450a94f113fd2c5a9979ac1ab9093021252436f60f2: Status 404 returned error can't find the container with id 4cf59c9455cbddd5f45fb450a94f113fd2c5a9979ac1ab9093021252436f60f2 Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.168968 4931 generic.go:334] "Generic (PLEG): container finished" podID="b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6" containerID="ad08a1031590ee8bb14fa8f169dd7a24b672d1e05913f033054dd83305aa291c" exitCode=0 Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.169047 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwvrh" event={"ID":"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6","Type":"ContainerDied","Data":"ad08a1031590ee8bb14fa8f169dd7a24b672d1e05913f033054dd83305aa291c"} Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.169316 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwvrh" event={"ID":"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6","Type":"ContainerStarted","Data":"4cf59c9455cbddd5f45fb450a94f113fd2c5a9979ac1ab9093021252436f60f2"} Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.172387 4931 generic.go:334] "Generic (PLEG): container finished" podID="c92e6619-370d-46ea-8925-a35dff600120" containerID="d16769b76161e3fcb7bced18f71db25b03fdcbd9d2f4da5bafb8f6ff00931ea0" exitCode=0 Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.172474 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zn6ld" event={"ID":"c92e6619-370d-46ea-8925-a35dff600120","Type":"ContainerDied","Data":"d16769b76161e3fcb7bced18f71db25b03fdcbd9d2f4da5bafb8f6ff00931ea0"} Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.172523 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zn6ld" event={"ID":"c92e6619-370d-46ea-8925-a35dff600120","Type":"ContainerStarted","Data":"1be0022430fb53125cce4ab6855e753476ccebd1577f2b02013abbc3e085f2a6"} Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.586099 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-56f9c54474-44w8q"] Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.586275 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" podUID="f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6" containerName="controller-manager" containerID="cri-o://f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc" gracePeriod=30 Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.630596 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb"] Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.630846 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" podUID="1fa096fe-9cff-4481-8e50-6eeb942b2354" containerName="route-controller-manager" containerID="cri-o://43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5" gracePeriod=30 Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.822991 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.887359 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2jpnh"] Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.888728 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.890904 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 29 00:05:14 crc kubenswrapper[4931]: I1129 00:05:14.896276 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2jpnh"] Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.073022 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59f58c30-f7fe-4141-ab70-122e133cccdc-catalog-content\") pod \"certified-operators-2jpnh\" (UID: \"59f58c30-f7fe-4141-ab70-122e133cccdc\") " pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.074130 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59f58c30-f7fe-4141-ab70-122e133cccdc-utilities\") pod \"certified-operators-2jpnh\" (UID: \"59f58c30-f7fe-4141-ab70-122e133cccdc\") " pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.074166 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75vcr\" (UniqueName: \"kubernetes.io/projected/59f58c30-f7fe-4141-ab70-122e133cccdc-kube-api-access-75vcr\") pod \"certified-operators-2jpnh\" (UID: \"59f58c30-f7fe-4141-ab70-122e133cccdc\") " pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.089581 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-28xxs"] Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.093332 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.093576 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fa096fe-9cff-4481-8e50-6eeb942b2354" containerName="route-controller-manager" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.094041 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.094688 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.096610 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.098224 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-28xxs"] Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.177548 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59f58c30-f7fe-4141-ab70-122e133cccdc-catalog-content\") pod \"certified-operators-2jpnh\" (UID: \"59f58c30-f7fe-4141-ab70-122e133cccdc\") " pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.177656 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59f58c30-f7fe-4141-ab70-122e133cccdc-utilities\") pod \"certified-operators-2jpnh\" (UID: \"59f58c30-f7fe-4141-ab70-122e133cccdc\") " pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.177691 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75vcr\" (UniqueName: \"kubernetes.io/projected/59f58c30-f7fe-4141-ab70-122e133cccdc-kube-api-access-75vcr\") pod \"certified-operators-2jpnh\" (UID: \"59f58c30-f7fe-4141-ab70-122e133cccdc\") " pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.178082 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59f58c30-f7fe-4141-ab70-122e133cccdc-catalog-content\") pod \"certified-operators-2jpnh\" (UID: \"59f58c30-f7fe-4141-ab70-122e133cccdc\") " pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.178120 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59f58c30-f7fe-4141-ab70-122e133cccdc-utilities\") pod \"certified-operators-2jpnh\" (UID: \"59f58c30-f7fe-4141-ab70-122e133cccdc\") " pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.179794 4931 generic.go:334] "Generic (PLEG): container finished" podID="f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6" containerID="f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc" exitCode=0 Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.179853 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" event={"ID":"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6","Type":"ContainerDied","Data":"f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc"} Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.179890 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.179920 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56f9c54474-44w8q" event={"ID":"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6","Type":"ContainerDied","Data":"7ecabf962d43ad58ded13d9b581178d36c54e2f759f5e2a4f3861ecb1d53dec7"} Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.179942 4931 scope.go:117] "RemoveContainer" containerID="f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.191526 4931 generic.go:334] "Generic (PLEG): container finished" podID="1fa096fe-9cff-4481-8e50-6eeb942b2354" containerID="43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5" exitCode=0 Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.191602 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.191601 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" event={"ID":"1fa096fe-9cff-4481-8e50-6eeb942b2354","Type":"ContainerDied","Data":"43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5"} Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.191649 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb" event={"ID":"1fa096fe-9cff-4481-8e50-6eeb942b2354","Type":"ContainerDied","Data":"8cc4320458669171156861a22cb46c73eba169c24cbf5883f3c4a34ebb49f206"} Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.197776 4931 generic.go:334] "Generic (PLEG): container finished" podID="b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6" containerID="4e8ead588e328de43ab8a695836706db355ff34db077be682b8d6dd667deda86" exitCode=0 Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.197828 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwvrh" event={"ID":"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6","Type":"ContainerDied","Data":"4e8ead588e328de43ab8a695836706db355ff34db077be682b8d6dd667deda86"} Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.198559 4931 scope.go:117] "RemoveContainer" containerID="f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.198561 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75vcr\" (UniqueName: \"kubernetes.io/projected/59f58c30-f7fe-4141-ab70-122e133cccdc-kube-api-access-75vcr\") pod \"certified-operators-2jpnh\" (UID: \"59f58c30-f7fe-4141-ab70-122e133cccdc\") " pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:15 crc kubenswrapper[4931]: E1129 00:05:15.198931 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc\": container with ID starting with f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc not found: ID does not exist" containerID="f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.198983 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc"} err="failed to get container status \"f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc\": rpc error: code = NotFound desc = could not find container \"f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc\": container with ID starting with f0a1c4cac53239753ab9ff4f39d946594452fb44f60b01d73dcff7f7a1acdcfc not found: ID does not exist" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.199006 4931 scope.go:117] "RemoveContainer" containerID="43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.210752 4931 scope.go:117] "RemoveContainer" containerID="43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5" Nov 29 00:05:15 crc kubenswrapper[4931]: E1129 00:05:15.211272 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5\": container with ID starting with 43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5 not found: ID does not exist" containerID="43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.211339 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5"} err="failed to get container status \"43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5\": rpc error: code = NotFound desc = could not find container \"43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5\": container with ID starting with 43d6741060a7be50bd43bfd1de5abe383d3badfebe7e9d72fe4f8360a238b1f5 not found: ID does not exist" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.221356 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.280168 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h75xg\" (UniqueName: \"kubernetes.io/projected/1fa096fe-9cff-4481-8e50-6eeb942b2354-kube-api-access-h75xg\") pod \"1fa096fe-9cff-4481-8e50-6eeb942b2354\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.280221 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1fa096fe-9cff-4481-8e50-6eeb942b2354-client-ca\") pod \"1fa096fe-9cff-4481-8e50-6eeb942b2354\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.280261 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-config\") pod \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.280307 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-serving-cert\") pod \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.280358 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-client-ca\") pod \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.280378 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fa096fe-9cff-4481-8e50-6eeb942b2354-serving-cert\") pod \"1fa096fe-9cff-4481-8e50-6eeb942b2354\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.280425 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa096fe-9cff-4481-8e50-6eeb942b2354-config\") pod \"1fa096fe-9cff-4481-8e50-6eeb942b2354\" (UID: \"1fa096fe-9cff-4481-8e50-6eeb942b2354\") " Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.280464 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsr7l\" (UniqueName: \"kubernetes.io/projected/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-kube-api-access-tsr7l\") pod \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.280506 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-proxy-ca-bundles\") pod \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\" (UID: \"f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6\") " Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.280660 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c31367fe-1dab-47be-906b-9324307948d0-catalog-content\") pod \"redhat-operators-28xxs\" (UID: \"c31367fe-1dab-47be-906b-9324307948d0\") " pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.280712 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc9pp\" (UniqueName: \"kubernetes.io/projected/c31367fe-1dab-47be-906b-9324307948d0-kube-api-access-rc9pp\") pod \"redhat-operators-28xxs\" (UID: \"c31367fe-1dab-47be-906b-9324307948d0\") " pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.280736 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c31367fe-1dab-47be-906b-9324307948d0-utilities\") pod \"redhat-operators-28xxs\" (UID: \"c31367fe-1dab-47be-906b-9324307948d0\") " pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.282038 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fa096fe-9cff-4481-8e50-6eeb942b2354-config" (OuterVolumeSpecName: "config") pod "1fa096fe-9cff-4481-8e50-6eeb942b2354" (UID: "1fa096fe-9cff-4481-8e50-6eeb942b2354"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.282069 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fa096fe-9cff-4481-8e50-6eeb942b2354-client-ca" (OuterVolumeSpecName: "client-ca") pod "1fa096fe-9cff-4481-8e50-6eeb942b2354" (UID: "1fa096fe-9cff-4481-8e50-6eeb942b2354"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.283093 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-client-ca" (OuterVolumeSpecName: "client-ca") pod "f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6" (UID: "f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.283360 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-config" (OuterVolumeSpecName: "config") pod "f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6" (UID: "f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.283450 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6" (UID: "f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.286118 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-kube-api-access-tsr7l" (OuterVolumeSpecName: "kube-api-access-tsr7l") pod "f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6" (UID: "f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6"). InnerVolumeSpecName "kube-api-access-tsr7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.286644 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fa096fe-9cff-4481-8e50-6eeb942b2354-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1fa096fe-9cff-4481-8e50-6eeb942b2354" (UID: "1fa096fe-9cff-4481-8e50-6eeb942b2354"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.286830 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6" (UID: "f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.287028 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fa096fe-9cff-4481-8e50-6eeb942b2354-kube-api-access-h75xg" (OuterVolumeSpecName: "kube-api-access-h75xg") pod "1fa096fe-9cff-4481-8e50-6eeb942b2354" (UID: "1fa096fe-9cff-4481-8e50-6eeb942b2354"). InnerVolumeSpecName "kube-api-access-h75xg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382160 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c31367fe-1dab-47be-906b-9324307948d0-catalog-content\") pod \"redhat-operators-28xxs\" (UID: \"c31367fe-1dab-47be-906b-9324307948d0\") " pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382517 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc9pp\" (UniqueName: \"kubernetes.io/projected/c31367fe-1dab-47be-906b-9324307948d0-kube-api-access-rc9pp\") pod \"redhat-operators-28xxs\" (UID: \"c31367fe-1dab-47be-906b-9324307948d0\") " pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382539 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c31367fe-1dab-47be-906b-9324307948d0-utilities\") pod \"redhat-operators-28xxs\" (UID: \"c31367fe-1dab-47be-906b-9324307948d0\") " pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382586 4931 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382600 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fa096fe-9cff-4481-8e50-6eeb942b2354-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382612 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa096fe-9cff-4481-8e50-6eeb942b2354-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382623 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsr7l\" (UniqueName: \"kubernetes.io/projected/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-kube-api-access-tsr7l\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382636 4931 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382646 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h75xg\" (UniqueName: \"kubernetes.io/projected/1fa096fe-9cff-4481-8e50-6eeb942b2354-kube-api-access-h75xg\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382656 4931 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1fa096fe-9cff-4481-8e50-6eeb942b2354-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382667 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382678 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.382990 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c31367fe-1dab-47be-906b-9324307948d0-catalog-content\") pod \"redhat-operators-28xxs\" (UID: \"c31367fe-1dab-47be-906b-9324307948d0\") " pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.383084 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c31367fe-1dab-47be-906b-9324307948d0-utilities\") pod \"redhat-operators-28xxs\" (UID: \"c31367fe-1dab-47be-906b-9324307948d0\") " pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.401747 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc9pp\" (UniqueName: \"kubernetes.io/projected/c31367fe-1dab-47be-906b-9324307948d0-kube-api-access-rc9pp\") pod \"redhat-operators-28xxs\" (UID: \"c31367fe-1dab-47be-906b-9324307948d0\") " pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.402207 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2jpnh"] Nov 29 00:05:15 crc kubenswrapper[4931]: W1129 00:05:15.406744 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59f58c30_f7fe_4141_ab70_122e133cccdc.slice/crio-7a49c0df9d2e09ff7322089055954337d821b48ddc4ed2dee825d55e5055d1f5 WatchSource:0}: Error finding container 7a49c0df9d2e09ff7322089055954337d821b48ddc4ed2dee825d55e5055d1f5: Status 404 returned error can't find the container with id 7a49c0df9d2e09ff7322089055954337d821b48ddc4ed2dee825d55e5055d1f5 Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.411449 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.524566 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-56f9c54474-44w8q"] Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.530144 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-56f9c54474-44w8q"] Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.539680 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb"] Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.553512 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5b6b6f74-rt9mb"] Nov 29 00:05:15 crc kubenswrapper[4931]: I1129 00:05:15.817323 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-28xxs"] Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.207941 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwvrh" event={"ID":"b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6","Type":"ContainerStarted","Data":"812b3acedc1149c689945c9f421712817a1b53931fc493877868e52a7eb723ca"} Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.210154 4931 generic.go:334] "Generic (PLEG): container finished" podID="c92e6619-370d-46ea-8925-a35dff600120" containerID="0549faedaf7419967754addc420539737620876e84f6bbe303626260f1f73594" exitCode=0 Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.210226 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zn6ld" event={"ID":"c92e6619-370d-46ea-8925-a35dff600120","Type":"ContainerDied","Data":"0549faedaf7419967754addc420539737620876e84f6bbe303626260f1f73594"} Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.213759 4931 generic.go:334] "Generic (PLEG): container finished" podID="c31367fe-1dab-47be-906b-9324307948d0" containerID="a476dd40e9992628b418535c21a0912043b47e851363acc7d68cc6b1d405defd" exitCode=0 Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.213864 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28xxs" event={"ID":"c31367fe-1dab-47be-906b-9324307948d0","Type":"ContainerDied","Data":"a476dd40e9992628b418535c21a0912043b47e851363acc7d68cc6b1d405defd"} Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.213914 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28xxs" event={"ID":"c31367fe-1dab-47be-906b-9324307948d0","Type":"ContainerStarted","Data":"c3fc40b39b554893277ab5aec967824e925b460d63e2ad5634de1e21f9ba3d02"} Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.220057 4931 generic.go:334] "Generic (PLEG): container finished" podID="59f58c30-f7fe-4141-ab70-122e133cccdc" containerID="7f42064d17eef14c72ba5fa14d986fe9d400d78de8d8cfff0c65bcc45530dedd" exitCode=0 Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.220133 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jpnh" event={"ID":"59f58c30-f7fe-4141-ab70-122e133cccdc","Type":"ContainerDied","Data":"7f42064d17eef14c72ba5fa14d986fe9d400d78de8d8cfff0c65bcc45530dedd"} Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.220167 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jpnh" event={"ID":"59f58c30-f7fe-4141-ab70-122e133cccdc","Type":"ContainerStarted","Data":"7a49c0df9d2e09ff7322089055954337d821b48ddc4ed2dee825d55e5055d1f5"} Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.230581 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jwvrh" podStartSLOduration=2.744978253 podStartE2EDuration="4.230564266s" podCreationTimestamp="2025-11-29 00:05:12 +0000 UTC" firstStartedPulling="2025-11-29 00:05:14.170908689 +0000 UTC m=+327.332801921" lastFinishedPulling="2025-11-29 00:05:15.656494702 +0000 UTC m=+328.818387934" observedRunningTime="2025-11-29 00:05:16.229608349 +0000 UTC m=+329.391501661" watchObservedRunningTime="2025-11-29 00:05:16.230564266 +0000 UTC m=+329.392457498" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.373016 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.492899 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-565f96c579-54bkv"] Nov 29 00:05:16 crc kubenswrapper[4931]: E1129 00:05:16.493213 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fa096fe-9cff-4481-8e50-6eeb942b2354" containerName="route-controller-manager" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.493232 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fa096fe-9cff-4481-8e50-6eeb942b2354" containerName="route-controller-manager" Nov 29 00:05:16 crc kubenswrapper[4931]: E1129 00:05:16.493245 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6" containerName="controller-manager" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.493253 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6" containerName="controller-manager" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.493368 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6" containerName="controller-manager" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.493873 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.498195 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.498582 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.499393 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp"] Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.501223 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.502451 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.502919 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.503086 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.506650 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.506709 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.506749 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.506663 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.507163 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.507349 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.507578 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp"] Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.509290 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.516678 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-565f96c579-54bkv"] Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.517754 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.597090 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-config\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.597171 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-proxy-ca-bundles\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.597202 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-client-ca\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.597471 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-serving-cert\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.597494 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xjt2\" (UniqueName: \"kubernetes.io/projected/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-kube-api-access-5xjt2\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.698188 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-config\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.698238 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62lsj\" (UniqueName: \"kubernetes.io/projected/466a450f-3f73-4e72-a77f-b9dbe2efd186-kube-api-access-62lsj\") pod \"route-controller-manager-59988bbdc9-ptdcp\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.698262 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/466a450f-3f73-4e72-a77f-b9dbe2efd186-config\") pod \"route-controller-manager-59988bbdc9-ptdcp\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.698279 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/466a450f-3f73-4e72-a77f-b9dbe2efd186-serving-cert\") pod \"route-controller-manager-59988bbdc9-ptdcp\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.698299 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/466a450f-3f73-4e72-a77f-b9dbe2efd186-client-ca\") pod \"route-controller-manager-59988bbdc9-ptdcp\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.698368 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-proxy-ca-bundles\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.698396 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-client-ca\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.698423 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-serving-cert\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.698442 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xjt2\" (UniqueName: \"kubernetes.io/projected/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-kube-api-access-5xjt2\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.699368 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-proxy-ca-bundles\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.699650 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-client-ca\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.699722 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-config\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.710023 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-serving-cert\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.722457 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xjt2\" (UniqueName: \"kubernetes.io/projected/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-kube-api-access-5xjt2\") pod \"controller-manager-565f96c579-54bkv\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.799333 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62lsj\" (UniqueName: \"kubernetes.io/projected/466a450f-3f73-4e72-a77f-b9dbe2efd186-kube-api-access-62lsj\") pod \"route-controller-manager-59988bbdc9-ptdcp\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.799420 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/466a450f-3f73-4e72-a77f-b9dbe2efd186-config\") pod \"route-controller-manager-59988bbdc9-ptdcp\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.799468 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/466a450f-3f73-4e72-a77f-b9dbe2efd186-serving-cert\") pod \"route-controller-manager-59988bbdc9-ptdcp\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.799520 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/466a450f-3f73-4e72-a77f-b9dbe2efd186-client-ca\") pod \"route-controller-manager-59988bbdc9-ptdcp\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.800434 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/466a450f-3f73-4e72-a77f-b9dbe2efd186-config\") pod \"route-controller-manager-59988bbdc9-ptdcp\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.801050 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/466a450f-3f73-4e72-a77f-b9dbe2efd186-client-ca\") pod \"route-controller-manager-59988bbdc9-ptdcp\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.804650 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/466a450f-3f73-4e72-a77f-b9dbe2efd186-serving-cert\") pod \"route-controller-manager-59988bbdc9-ptdcp\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.818111 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.819420 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62lsj\" (UniqueName: \"kubernetes.io/projected/466a450f-3f73-4e72-a77f-b9dbe2efd186-kube-api-access-62lsj\") pod \"route-controller-manager-59988bbdc9-ptdcp\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:16 crc kubenswrapper[4931]: I1129 00:05:16.822935 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:17 crc kubenswrapper[4931]: I1129 00:05:17.223208 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fa096fe-9cff-4481-8e50-6eeb942b2354" path="/var/lib/kubelet/pods/1fa096fe-9cff-4481-8e50-6eeb942b2354/volumes" Nov 29 00:05:17 crc kubenswrapper[4931]: I1129 00:05:17.224344 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6" path="/var/lib/kubelet/pods/f6ed7a98-dd2a-4c13-8e70-ca4a5f0dbcb6/volumes" Nov 29 00:05:17 crc kubenswrapper[4931]: I1129 00:05:17.253854 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28xxs" event={"ID":"c31367fe-1dab-47be-906b-9324307948d0","Type":"ContainerStarted","Data":"b5af5dec7c2f70f7ac143f64fd4f4f64ec6710c3745bfdb03c67b3834f853e50"} Nov 29 00:05:17 crc kubenswrapper[4931]: I1129 00:05:17.256788 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jpnh" event={"ID":"59f58c30-f7fe-4141-ab70-122e133cccdc","Type":"ContainerStarted","Data":"1d4dadae7ba46d919333c87e686f8afeedf76bc6890d51ebaf31f80e1a91c16a"} Nov 29 00:05:17 crc kubenswrapper[4931]: I1129 00:05:17.260963 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zn6ld" event={"ID":"c92e6619-370d-46ea-8925-a35dff600120","Type":"ContainerStarted","Data":"1956a442515c9e6359d595ede6dd764b77ceba18bc1a0831e6a1e08e89a241a7"} Nov 29 00:05:17 crc kubenswrapper[4931]: I1129 00:05:17.290719 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-565f96c579-54bkv"] Nov 29 00:05:17 crc kubenswrapper[4931]: I1129 00:05:17.301594 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zn6ld" podStartSLOduration=2.524107412 podStartE2EDuration="5.301579626s" podCreationTimestamp="2025-11-29 00:05:12 +0000 UTC" firstStartedPulling="2025-11-29 00:05:14.17438141 +0000 UTC m=+327.336274642" lastFinishedPulling="2025-11-29 00:05:16.951853624 +0000 UTC m=+330.113746856" observedRunningTime="2025-11-29 00:05:17.299226998 +0000 UTC m=+330.461120240" watchObservedRunningTime="2025-11-29 00:05:17.301579626 +0000 UTC m=+330.463472858" Nov 29 00:05:17 crc kubenswrapper[4931]: I1129 00:05:17.318208 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp"] Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.267241 4931 generic.go:334] "Generic (PLEG): container finished" podID="59f58c30-f7fe-4141-ab70-122e133cccdc" containerID="1d4dadae7ba46d919333c87e686f8afeedf76bc6890d51ebaf31f80e1a91c16a" exitCode=0 Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.267468 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jpnh" event={"ID":"59f58c30-f7fe-4141-ab70-122e133cccdc","Type":"ContainerDied","Data":"1d4dadae7ba46d919333c87e686f8afeedf76bc6890d51ebaf31f80e1a91c16a"} Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.270147 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" event={"ID":"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7","Type":"ContainerStarted","Data":"083b135c5a48172b2897187736b72f03e9dc4b97a39f739429921c6759953fd3"} Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.270184 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" event={"ID":"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7","Type":"ContainerStarted","Data":"5408fd051bfe55439d34caed2aaefdf63fe8a7cb8511659c69bdc0c43c50503e"} Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.270342 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.273562 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" event={"ID":"466a450f-3f73-4e72-a77f-b9dbe2efd186","Type":"ContainerStarted","Data":"822fd440edd32e265c806cc70a65ca0a5dc26d749a8702e20214bf2fe9b80efc"} Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.273609 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" event={"ID":"466a450f-3f73-4e72-a77f-b9dbe2efd186","Type":"ContainerStarted","Data":"5c0f4245cc9143e3638a76b1b446221de77002123d003879db4a16ecf03a6fd3"} Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.274163 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.276052 4931 generic.go:334] "Generic (PLEG): container finished" podID="c31367fe-1dab-47be-906b-9324307948d0" containerID="b5af5dec7c2f70f7ac143f64fd4f4f64ec6710c3745bfdb03c67b3834f853e50" exitCode=0 Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.276848 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28xxs" event={"ID":"c31367fe-1dab-47be-906b-9324307948d0","Type":"ContainerDied","Data":"b5af5dec7c2f70f7ac143f64fd4f4f64ec6710c3745bfdb03c67b3834f853e50"} Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.279152 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.280988 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.324141 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" podStartSLOduration=4.324126849 podStartE2EDuration="4.324126849s" podCreationTimestamp="2025-11-29 00:05:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:05:18.30416128 +0000 UTC m=+331.466054522" watchObservedRunningTime="2025-11-29 00:05:18.324126849 +0000 UTC m=+331.486020081" Nov 29 00:05:18 crc kubenswrapper[4931]: I1129 00:05:18.324973 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" podStartSLOduration=4.324967084 podStartE2EDuration="4.324967084s" podCreationTimestamp="2025-11-29 00:05:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:05:18.320985468 +0000 UTC m=+331.482878710" watchObservedRunningTime="2025-11-29 00:05:18.324967084 +0000 UTC m=+331.486860316" Nov 29 00:05:19 crc kubenswrapper[4931]: I1129 00:05:19.285486 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-28xxs" event={"ID":"c31367fe-1dab-47be-906b-9324307948d0","Type":"ContainerStarted","Data":"b0ca460595153d94c91585e0edbd1c7bdd670477551f41120a560abd535ba96b"} Nov 29 00:05:19 crc kubenswrapper[4931]: I1129 00:05:19.289681 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jpnh" event={"ID":"59f58c30-f7fe-4141-ab70-122e133cccdc","Type":"ContainerStarted","Data":"82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf"} Nov 29 00:05:19 crc kubenswrapper[4931]: I1129 00:05:19.319254 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-28xxs" podStartSLOduration=1.7084119869999999 podStartE2EDuration="4.319229205s" podCreationTimestamp="2025-11-29 00:05:15 +0000 UTC" firstStartedPulling="2025-11-29 00:05:16.217964191 +0000 UTC m=+329.379857423" lastFinishedPulling="2025-11-29 00:05:18.828781409 +0000 UTC m=+331.990674641" observedRunningTime="2025-11-29 00:05:19.313004914 +0000 UTC m=+332.474898256" watchObservedRunningTime="2025-11-29 00:05:19.319229205 +0000 UTC m=+332.481122467" Nov 29 00:05:19 crc kubenswrapper[4931]: I1129 00:05:19.344319 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2jpnh" podStartSLOduration=2.672233406 podStartE2EDuration="5.344302643s" podCreationTimestamp="2025-11-29 00:05:14 +0000 UTC" firstStartedPulling="2025-11-29 00:05:16.221259926 +0000 UTC m=+329.383153198" lastFinishedPulling="2025-11-29 00:05:18.893329213 +0000 UTC m=+332.055222435" observedRunningTime="2025-11-29 00:05:19.339816622 +0000 UTC m=+332.501709854" watchObservedRunningTime="2025-11-29 00:05:19.344302643 +0000 UTC m=+332.506195875" Nov 29 00:05:22 crc kubenswrapper[4931]: I1129 00:05:22.844529 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:22 crc kubenswrapper[4931]: I1129 00:05:22.848048 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:22 crc kubenswrapper[4931]: I1129 00:05:22.931723 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:23 crc kubenswrapper[4931]: I1129 00:05:23.003873 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:23 crc kubenswrapper[4931]: I1129 00:05:23.003935 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:23 crc kubenswrapper[4931]: I1129 00:05:23.043381 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:23 crc kubenswrapper[4931]: I1129 00:05:23.392093 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jwvrh" Nov 29 00:05:23 crc kubenswrapper[4931]: I1129 00:05:23.404919 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zn6ld" Nov 29 00:05:25 crc kubenswrapper[4931]: I1129 00:05:25.222350 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:25 crc kubenswrapper[4931]: I1129 00:05:25.222433 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:25 crc kubenswrapper[4931]: I1129 00:05:25.299713 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:25 crc kubenswrapper[4931]: I1129 00:05:25.400066 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 00:05:25 crc kubenswrapper[4931]: I1129 00:05:25.412200 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:25 crc kubenswrapper[4931]: I1129 00:05:25.412296 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:25 crc kubenswrapper[4931]: I1129 00:05:25.464658 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:26 crc kubenswrapper[4931]: I1129 00:05:26.414419 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-28xxs" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.306417 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2l4m5"] Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.307626 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.331029 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2l4m5"] Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.457717 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42340b86-9c5b-4906-bae0-1fbc76c31a6c-registry-certificates\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.457788 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42340b86-9c5b-4906-bae0-1fbc76c31a6c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.457831 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42340b86-9c5b-4906-bae0-1fbc76c31a6c-registry-tls\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.458000 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.458036 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42340b86-9c5b-4906-bae0-1fbc76c31a6c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.458080 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42340b86-9c5b-4906-bae0-1fbc76c31a6c-bound-sa-token\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.458111 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hqt7\" (UniqueName: \"kubernetes.io/projected/42340b86-9c5b-4906-bae0-1fbc76c31a6c-kube-api-access-7hqt7\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.458188 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42340b86-9c5b-4906-bae0-1fbc76c31a6c-trusted-ca\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.489627 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.559568 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42340b86-9c5b-4906-bae0-1fbc76c31a6c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.559624 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42340b86-9c5b-4906-bae0-1fbc76c31a6c-registry-tls\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.559682 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42340b86-9c5b-4906-bae0-1fbc76c31a6c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.559705 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42340b86-9c5b-4906-bae0-1fbc76c31a6c-bound-sa-token\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.559733 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hqt7\" (UniqueName: \"kubernetes.io/projected/42340b86-9c5b-4906-bae0-1fbc76c31a6c-kube-api-access-7hqt7\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.559760 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42340b86-9c5b-4906-bae0-1fbc76c31a6c-trusted-ca\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.559819 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42340b86-9c5b-4906-bae0-1fbc76c31a6c-registry-certificates\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.560722 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42340b86-9c5b-4906-bae0-1fbc76c31a6c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.561422 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42340b86-9c5b-4906-bae0-1fbc76c31a6c-registry-certificates\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.562941 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42340b86-9c5b-4906-bae0-1fbc76c31a6c-trusted-ca\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.566646 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42340b86-9c5b-4906-bae0-1fbc76c31a6c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.567646 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42340b86-9c5b-4906-bae0-1fbc76c31a6c-registry-tls\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.575920 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42340b86-9c5b-4906-bae0-1fbc76c31a6c-bound-sa-token\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.576108 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hqt7\" (UniqueName: \"kubernetes.io/projected/42340b86-9c5b-4906-bae0-1fbc76c31a6c-kube-api-access-7hqt7\") pod \"image-registry-66df7c8f76-2l4m5\" (UID: \"42340b86-9c5b-4906-bae0-1fbc76c31a6c\") " pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:35 crc kubenswrapper[4931]: I1129 00:05:35.621529 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:36 crc kubenswrapper[4931]: I1129 00:05:36.098467 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2l4m5"] Nov 29 00:05:36 crc kubenswrapper[4931]: W1129 00:05:36.104569 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42340b86_9c5b_4906_bae0_1fbc76c31a6c.slice/crio-04a0686b92e5c5aee198529e4a9e4ef5cff8e1fde2a68268121653ce0c55bf3f WatchSource:0}: Error finding container 04a0686b92e5c5aee198529e4a9e4ef5cff8e1fde2a68268121653ce0c55bf3f: Status 404 returned error can't find the container with id 04a0686b92e5c5aee198529e4a9e4ef5cff8e1fde2a68268121653ce0c55bf3f Nov 29 00:05:36 crc kubenswrapper[4931]: I1129 00:05:36.432557 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" event={"ID":"42340b86-9c5b-4906-bae0-1fbc76c31a6c","Type":"ContainerStarted","Data":"04a0686b92e5c5aee198529e4a9e4ef5cff8e1fde2a68268121653ce0c55bf3f"} Nov 29 00:05:42 crc kubenswrapper[4931]: I1129 00:05:42.477689 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" event={"ID":"42340b86-9c5b-4906-bae0-1fbc76c31a6c","Type":"ContainerStarted","Data":"de05343409afa46e6653016eeacc559eb74a41f05b2dbe429f77e89f8cd5bee8"} Nov 29 00:05:43 crc kubenswrapper[4931]: I1129 00:05:43.482122 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:05:43 crc kubenswrapper[4931]: I1129 00:05:43.507418 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" podStartSLOduration=8.507400494 podStartE2EDuration="8.507400494s" podCreationTimestamp="2025-11-29 00:05:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:05:43.50314803 +0000 UTC m=+356.665041272" watchObservedRunningTime="2025-11-29 00:05:43.507400494 +0000 UTC m=+356.669293726" Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.201836 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-565f96c579-54bkv"] Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.202489 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" podUID="4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7" containerName="controller-manager" containerID="cri-o://083b135c5a48172b2897187736b72f03e9dc4b97a39f739429921c6759953fd3" gracePeriod=30 Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.501656 4931 generic.go:334] "Generic (PLEG): container finished" podID="4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7" containerID="083b135c5a48172b2897187736b72f03e9dc4b97a39f739429921c6759953fd3" exitCode=0 Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.501752 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" event={"ID":"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7","Type":"ContainerDied","Data":"083b135c5a48172b2897187736b72f03e9dc4b97a39f739429921c6759953fd3"} Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.791050 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.910037 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-client-ca\") pod \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.910142 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xjt2\" (UniqueName: \"kubernetes.io/projected/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-kube-api-access-5xjt2\") pod \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.910183 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-proxy-ca-bundles\") pod \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.910247 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-serving-cert\") pod \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.910286 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-config\") pod \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\" (UID: \"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7\") " Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.910932 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-client-ca" (OuterVolumeSpecName: "client-ca") pod "4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7" (UID: "4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.910950 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7" (UID: "4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.911254 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-config" (OuterVolumeSpecName: "config") pod "4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7" (UID: "4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.915229 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7" (UID: "4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:05:46 crc kubenswrapper[4931]: I1129 00:05:46.916451 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-kube-api-access-5xjt2" (OuterVolumeSpecName: "kube-api-access-5xjt2") pod "4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7" (UID: "4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7"). InnerVolumeSpecName "kube-api-access-5xjt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.011387 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xjt2\" (UniqueName: \"kubernetes.io/projected/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-kube-api-access-5xjt2\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.011421 4931 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.011435 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.011449 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.011460 4931 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.513959 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-67559f8cc6-mz5rl"] Nov 29 00:05:47 crc kubenswrapper[4931]: E1129 00:05:47.515963 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7" containerName="controller-manager" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.516303 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7" containerName="controller-manager" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.520709 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7" containerName="controller-manager" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.521614 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.521969 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" event={"ID":"4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7","Type":"ContainerDied","Data":"5408fd051bfe55439d34caed2aaefdf63fe8a7cb8511659c69bdc0c43c50503e"} Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.522056 4931 scope.go:117] "RemoveContainer" containerID="083b135c5a48172b2897187736b72f03e9dc4b97a39f739429921c6759953fd3" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.522117 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-565f96c579-54bkv" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.532683 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-67559f8cc6-mz5rl"] Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.582110 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-565f96c579-54bkv"] Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.589538 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-565f96c579-54bkv"] Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.620216 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-proxy-ca-bundles\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.620287 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-serving-cert\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.620331 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wcxg\" (UniqueName: \"kubernetes.io/projected/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-kube-api-access-6wcxg\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.620392 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-client-ca\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.620427 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-config\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.722753 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-serving-cert\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.722896 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wcxg\" (UniqueName: \"kubernetes.io/projected/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-kube-api-access-6wcxg\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.722977 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-client-ca\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.723024 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-config\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.723085 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-proxy-ca-bundles\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.723756 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-client-ca\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.724540 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-config\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.725929 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-proxy-ca-bundles\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.728221 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-serving-cert\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.747242 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wcxg\" (UniqueName: \"kubernetes.io/projected/252ac9c2-719d-4fa3-8b5f-bf2390e2172f-kube-api-access-6wcxg\") pod \"controller-manager-67559f8cc6-mz5rl\" (UID: \"252ac9c2-719d-4fa3-8b5f-bf2390e2172f\") " pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:47 crc kubenswrapper[4931]: I1129 00:05:47.886871 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:48 crc kubenswrapper[4931]: I1129 00:05:48.150149 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-67559f8cc6-mz5rl"] Nov 29 00:05:48 crc kubenswrapper[4931]: W1129 00:05:48.154223 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod252ac9c2_719d_4fa3_8b5f_bf2390e2172f.slice/crio-46becbf8ef44a5b411527922f48edeaef2f46a2d08d5951de7fcaa660bef293f WatchSource:0}: Error finding container 46becbf8ef44a5b411527922f48edeaef2f46a2d08d5951de7fcaa660bef293f: Status 404 returned error can't find the container with id 46becbf8ef44a5b411527922f48edeaef2f46a2d08d5951de7fcaa660bef293f Nov 29 00:05:48 crc kubenswrapper[4931]: I1129 00:05:48.530625 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" event={"ID":"252ac9c2-719d-4fa3-8b5f-bf2390e2172f","Type":"ContainerStarted","Data":"3ba3fb7a180fb8ee60b7da337dcef8ab6c08521c6797ef03936b4f2ba8c3b939"} Nov 29 00:05:48 crc kubenswrapper[4931]: I1129 00:05:48.531029 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" event={"ID":"252ac9c2-719d-4fa3-8b5f-bf2390e2172f","Type":"ContainerStarted","Data":"46becbf8ef44a5b411527922f48edeaef2f46a2d08d5951de7fcaa660bef293f"} Nov 29 00:05:48 crc kubenswrapper[4931]: I1129 00:05:48.531064 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:48 crc kubenswrapper[4931]: I1129 00:05:48.547478 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" Nov 29 00:05:48 crc kubenswrapper[4931]: I1129 00:05:48.557033 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-67559f8cc6-mz5rl" podStartSLOduration=2.557016364 podStartE2EDuration="2.557016364s" podCreationTimestamp="2025-11-29 00:05:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:05:48.555690366 +0000 UTC m=+361.717583618" watchObservedRunningTime="2025-11-29 00:05:48.557016364 +0000 UTC m=+361.718909596" Nov 29 00:05:49 crc kubenswrapper[4931]: I1129 00:05:49.220558 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7" path="/var/lib/kubelet/pods/4e2299ce-5cff-4c71-b4d1-03ad7c5d2fc7/volumes" Nov 29 00:05:53 crc kubenswrapper[4931]: I1129 00:05:53.735088 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:05:53 crc kubenswrapper[4931]: I1129 00:05:53.735713 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:06:05 crc kubenswrapper[4931]: I1129 00:06:05.627037 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-2l4m5" Nov 29 00:06:05 crc kubenswrapper[4931]: I1129 00:06:05.704639 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8j97r"] Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.210692 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp"] Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.210977 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" podUID="466a450f-3f73-4e72-a77f-b9dbe2efd186" containerName="route-controller-manager" containerID="cri-o://822fd440edd32e265c806cc70a65ca0a5dc26d749a8702e20214bf2fe9b80efc" gracePeriod=30 Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.670861 4931 generic.go:334] "Generic (PLEG): container finished" podID="466a450f-3f73-4e72-a77f-b9dbe2efd186" containerID="822fd440edd32e265c806cc70a65ca0a5dc26d749a8702e20214bf2fe9b80efc" exitCode=0 Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.670947 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" event={"ID":"466a450f-3f73-4e72-a77f-b9dbe2efd186","Type":"ContainerDied","Data":"822fd440edd32e265c806cc70a65ca0a5dc26d749a8702e20214bf2fe9b80efc"} Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.780705 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.843564 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62lsj\" (UniqueName: \"kubernetes.io/projected/466a450f-3f73-4e72-a77f-b9dbe2efd186-kube-api-access-62lsj\") pod \"466a450f-3f73-4e72-a77f-b9dbe2efd186\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.843708 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/466a450f-3f73-4e72-a77f-b9dbe2efd186-serving-cert\") pod \"466a450f-3f73-4e72-a77f-b9dbe2efd186\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.843756 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/466a450f-3f73-4e72-a77f-b9dbe2efd186-config\") pod \"466a450f-3f73-4e72-a77f-b9dbe2efd186\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.843848 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/466a450f-3f73-4e72-a77f-b9dbe2efd186-client-ca\") pod \"466a450f-3f73-4e72-a77f-b9dbe2efd186\" (UID: \"466a450f-3f73-4e72-a77f-b9dbe2efd186\") " Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.845489 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/466a450f-3f73-4e72-a77f-b9dbe2efd186-client-ca" (OuterVolumeSpecName: "client-ca") pod "466a450f-3f73-4e72-a77f-b9dbe2efd186" (UID: "466a450f-3f73-4e72-a77f-b9dbe2efd186"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.846750 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/466a450f-3f73-4e72-a77f-b9dbe2efd186-config" (OuterVolumeSpecName: "config") pod "466a450f-3f73-4e72-a77f-b9dbe2efd186" (UID: "466a450f-3f73-4e72-a77f-b9dbe2efd186"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.854288 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/466a450f-3f73-4e72-a77f-b9dbe2efd186-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "466a450f-3f73-4e72-a77f-b9dbe2efd186" (UID: "466a450f-3f73-4e72-a77f-b9dbe2efd186"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.854927 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/466a450f-3f73-4e72-a77f-b9dbe2efd186-kube-api-access-62lsj" (OuterVolumeSpecName: "kube-api-access-62lsj") pod "466a450f-3f73-4e72-a77f-b9dbe2efd186" (UID: "466a450f-3f73-4e72-a77f-b9dbe2efd186"). InnerVolumeSpecName "kube-api-access-62lsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.946173 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62lsj\" (UniqueName: \"kubernetes.io/projected/466a450f-3f73-4e72-a77f-b9dbe2efd186-kube-api-access-62lsj\") on node \"crc\" DevicePath \"\"" Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.946229 4931 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/466a450f-3f73-4e72-a77f-b9dbe2efd186-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.946249 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/466a450f-3f73-4e72-a77f-b9dbe2efd186-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:06:06 crc kubenswrapper[4931]: I1129 00:06:06.946265 4931 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/466a450f-3f73-4e72-a77f-b9dbe2efd186-client-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.528393 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh"] Nov 29 00:06:07 crc kubenswrapper[4931]: E1129 00:06:07.528652 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="466a450f-3f73-4e72-a77f-b9dbe2efd186" containerName="route-controller-manager" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.528667 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="466a450f-3f73-4e72-a77f-b9dbe2efd186" containerName="route-controller-manager" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.528793 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="466a450f-3f73-4e72-a77f-b9dbe2efd186" containerName="route-controller-manager" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.529409 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.546525 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh"] Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.657553 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48471d67-ae36-4dc4-978c-721b5fe54aef-client-ca\") pod \"route-controller-manager-cbfcf7998-s2hqh\" (UID: \"48471d67-ae36-4dc4-978c-721b5fe54aef\") " pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.657639 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48471d67-ae36-4dc4-978c-721b5fe54aef-config\") pod \"route-controller-manager-cbfcf7998-s2hqh\" (UID: \"48471d67-ae36-4dc4-978c-721b5fe54aef\") " pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.657681 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25lkv\" (UniqueName: \"kubernetes.io/projected/48471d67-ae36-4dc4-978c-721b5fe54aef-kube-api-access-25lkv\") pod \"route-controller-manager-cbfcf7998-s2hqh\" (UID: \"48471d67-ae36-4dc4-978c-721b5fe54aef\") " pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.657705 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48471d67-ae36-4dc4-978c-721b5fe54aef-serving-cert\") pod \"route-controller-manager-cbfcf7998-s2hqh\" (UID: \"48471d67-ae36-4dc4-978c-721b5fe54aef\") " pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.679337 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" event={"ID":"466a450f-3f73-4e72-a77f-b9dbe2efd186","Type":"ContainerDied","Data":"5c0f4245cc9143e3638a76b1b446221de77002123d003879db4a16ecf03a6fd3"} Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.679386 4931 scope.go:117] "RemoveContainer" containerID="822fd440edd32e265c806cc70a65ca0a5dc26d749a8702e20214bf2fe9b80efc" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.679448 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.704473 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp"] Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.709373 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59988bbdc9-ptdcp"] Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.758517 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25lkv\" (UniqueName: \"kubernetes.io/projected/48471d67-ae36-4dc4-978c-721b5fe54aef-kube-api-access-25lkv\") pod \"route-controller-manager-cbfcf7998-s2hqh\" (UID: \"48471d67-ae36-4dc4-978c-721b5fe54aef\") " pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.758554 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48471d67-ae36-4dc4-978c-721b5fe54aef-serving-cert\") pod \"route-controller-manager-cbfcf7998-s2hqh\" (UID: \"48471d67-ae36-4dc4-978c-721b5fe54aef\") " pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.758600 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48471d67-ae36-4dc4-978c-721b5fe54aef-client-ca\") pod \"route-controller-manager-cbfcf7998-s2hqh\" (UID: \"48471d67-ae36-4dc4-978c-721b5fe54aef\") " pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.758650 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48471d67-ae36-4dc4-978c-721b5fe54aef-config\") pod \"route-controller-manager-cbfcf7998-s2hqh\" (UID: \"48471d67-ae36-4dc4-978c-721b5fe54aef\") " pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.759853 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48471d67-ae36-4dc4-978c-721b5fe54aef-config\") pod \"route-controller-manager-cbfcf7998-s2hqh\" (UID: \"48471d67-ae36-4dc4-978c-721b5fe54aef\") " pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.760155 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48471d67-ae36-4dc4-978c-721b5fe54aef-client-ca\") pod \"route-controller-manager-cbfcf7998-s2hqh\" (UID: \"48471d67-ae36-4dc4-978c-721b5fe54aef\") " pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.766317 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48471d67-ae36-4dc4-978c-721b5fe54aef-serving-cert\") pod \"route-controller-manager-cbfcf7998-s2hqh\" (UID: \"48471d67-ae36-4dc4-978c-721b5fe54aef\") " pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.778500 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25lkv\" (UniqueName: \"kubernetes.io/projected/48471d67-ae36-4dc4-978c-721b5fe54aef-kube-api-access-25lkv\") pod \"route-controller-manager-cbfcf7998-s2hqh\" (UID: \"48471d67-ae36-4dc4-978c-721b5fe54aef\") " pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:07 crc kubenswrapper[4931]: I1129 00:06:07.888721 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:08 crc kubenswrapper[4931]: I1129 00:06:08.308664 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh"] Nov 29 00:06:08 crc kubenswrapper[4931]: I1129 00:06:08.686416 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" event={"ID":"48471d67-ae36-4dc4-978c-721b5fe54aef","Type":"ContainerStarted","Data":"b9f011b2213b1267d500a975f2daf33bcea92985cb3cdafc96ab74ca763b0e3c"} Nov 29 00:06:08 crc kubenswrapper[4931]: I1129 00:06:08.686660 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:08 crc kubenswrapper[4931]: I1129 00:06:08.686671 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" event={"ID":"48471d67-ae36-4dc4-978c-721b5fe54aef","Type":"ContainerStarted","Data":"ddc2593bc2e08e52d8b142e697d69dc467de4d76458bfa22f836c607b397f280"} Nov 29 00:06:08 crc kubenswrapper[4931]: I1129 00:06:08.707080 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" podStartSLOduration=2.707057765 podStartE2EDuration="2.707057765s" podCreationTimestamp="2025-11-29 00:06:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:06:08.703873943 +0000 UTC m=+381.865767225" watchObservedRunningTime="2025-11-29 00:06:08.707057765 +0000 UTC m=+381.868951007" Nov 29 00:06:09 crc kubenswrapper[4931]: I1129 00:06:09.034588 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-cbfcf7998-s2hqh" Nov 29 00:06:09 crc kubenswrapper[4931]: I1129 00:06:09.219565 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="466a450f-3f73-4e72-a77f-b9dbe2efd186" path="/var/lib/kubelet/pods/466a450f-3f73-4e72-a77f-b9dbe2efd186/volumes" Nov 29 00:06:23 crc kubenswrapper[4931]: I1129 00:06:23.735300 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:06:23 crc kubenswrapper[4931]: I1129 00:06:23.735797 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:06:30 crc kubenswrapper[4931]: I1129 00:06:30.744192 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" podUID="93297fe9-2486-4b24-b6c6-cb917ab240c5" containerName="registry" containerID="cri-o://2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219" gracePeriod=30 Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.174286 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.318658 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-registry-tls\") pod \"93297fe9-2486-4b24-b6c6-cb917ab240c5\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.318763 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-bound-sa-token\") pod \"93297fe9-2486-4b24-b6c6-cb917ab240c5\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.318850 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/93297fe9-2486-4b24-b6c6-cb917ab240c5-installation-pull-secrets\") pod \"93297fe9-2486-4b24-b6c6-cb917ab240c5\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.318883 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgkqz\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-kube-api-access-kgkqz\") pod \"93297fe9-2486-4b24-b6c6-cb917ab240c5\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.318939 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/93297fe9-2486-4b24-b6c6-cb917ab240c5-registry-certificates\") pod \"93297fe9-2486-4b24-b6c6-cb917ab240c5\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.319270 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"93297fe9-2486-4b24-b6c6-cb917ab240c5\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.319334 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93297fe9-2486-4b24-b6c6-cb917ab240c5-trusted-ca\") pod \"93297fe9-2486-4b24-b6c6-cb917ab240c5\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.319365 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/93297fe9-2486-4b24-b6c6-cb917ab240c5-ca-trust-extracted\") pod \"93297fe9-2486-4b24-b6c6-cb917ab240c5\" (UID: \"93297fe9-2486-4b24-b6c6-cb917ab240c5\") " Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.329613 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93297fe9-2486-4b24-b6c6-cb917ab240c5-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "93297fe9-2486-4b24-b6c6-cb917ab240c5" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.330386 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93297fe9-2486-4b24-b6c6-cb917ab240c5-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "93297fe9-2486-4b24-b6c6-cb917ab240c5" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.336526 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "93297fe9-2486-4b24-b6c6-cb917ab240c5" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.338939 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-kube-api-access-kgkqz" (OuterVolumeSpecName: "kube-api-access-kgkqz") pod "93297fe9-2486-4b24-b6c6-cb917ab240c5" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5"). InnerVolumeSpecName "kube-api-access-kgkqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.340941 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93297fe9-2486-4b24-b6c6-cb917ab240c5-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "93297fe9-2486-4b24-b6c6-cb917ab240c5" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.342317 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93297fe9-2486-4b24-b6c6-cb917ab240c5-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "93297fe9-2486-4b24-b6c6-cb917ab240c5" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.346193 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "93297fe9-2486-4b24-b6c6-cb917ab240c5" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.346344 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "93297fe9-2486-4b24-b6c6-cb917ab240c5" (UID: "93297fe9-2486-4b24-b6c6-cb917ab240c5"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.420655 4931 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.420694 4931 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.420705 4931 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/93297fe9-2486-4b24-b6c6-cb917ab240c5-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.420716 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgkqz\" (UniqueName: \"kubernetes.io/projected/93297fe9-2486-4b24-b6c6-cb917ab240c5-kube-api-access-kgkqz\") on node \"crc\" DevicePath \"\"" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.420726 4931 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/93297fe9-2486-4b24-b6c6-cb917ab240c5-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.420736 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93297fe9-2486-4b24-b6c6-cb917ab240c5-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.420747 4931 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/93297fe9-2486-4b24-b6c6-cb917ab240c5-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.840255 4931 generic.go:334] "Generic (PLEG): container finished" podID="93297fe9-2486-4b24-b6c6-cb917ab240c5" containerID="2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219" exitCode=0 Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.840296 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" event={"ID":"93297fe9-2486-4b24-b6c6-cb917ab240c5","Type":"ContainerDied","Data":"2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219"} Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.840324 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.840346 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8j97r" event={"ID":"93297fe9-2486-4b24-b6c6-cb917ab240c5","Type":"ContainerDied","Data":"30f7807fcfda3e2816451654584376de3a71d6980df08546de1998da9ae3de0c"} Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.840371 4931 scope.go:117] "RemoveContainer" containerID="2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.855598 4931 scope.go:117] "RemoveContainer" containerID="2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219" Nov 29 00:06:31 crc kubenswrapper[4931]: E1129 00:06:31.856026 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219\": container with ID starting with 2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219 not found: ID does not exist" containerID="2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.856062 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219"} err="failed to get container status \"2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219\": rpc error: code = NotFound desc = could not find container \"2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219\": container with ID starting with 2b0c755429961003b77e89b919cad27da10c2583da288983770096337cacb219 not found: ID does not exist" Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.868319 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8j97r"] Nov 29 00:06:31 crc kubenswrapper[4931]: I1129 00:06:31.872928 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8j97r"] Nov 29 00:06:33 crc kubenswrapper[4931]: I1129 00:06:33.226293 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93297fe9-2486-4b24-b6c6-cb917ab240c5" path="/var/lib/kubelet/pods/93297fe9-2486-4b24-b6c6-cb917ab240c5/volumes" Nov 29 00:06:53 crc kubenswrapper[4931]: I1129 00:06:53.735086 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:06:53 crc kubenswrapper[4931]: I1129 00:06:53.735750 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:06:53 crc kubenswrapper[4931]: I1129 00:06:53.735823 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:06:53 crc kubenswrapper[4931]: I1129 00:06:53.736646 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"720e0a368eab3734412726774de172ffe278950e85b711a795977fe783122261"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 00:06:53 crc kubenswrapper[4931]: I1129 00:06:53.736758 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://720e0a368eab3734412726774de172ffe278950e85b711a795977fe783122261" gracePeriod=600 Nov 29 00:06:53 crc kubenswrapper[4931]: I1129 00:06:53.995859 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="720e0a368eab3734412726774de172ffe278950e85b711a795977fe783122261" exitCode=0 Nov 29 00:06:53 crc kubenswrapper[4931]: I1129 00:06:53.996020 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"720e0a368eab3734412726774de172ffe278950e85b711a795977fe783122261"} Nov 29 00:06:53 crc kubenswrapper[4931]: I1129 00:06:53.996299 4931 scope.go:117] "RemoveContainer" containerID="980ed516ffd7c1fb5194849bb8551bb908435b4be7b0b3b4e5e5a5b21a9e5db9" Nov 29 00:06:55 crc kubenswrapper[4931]: I1129 00:06:55.003066 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"ffd259621969111fa5cf386e3c4f80aaf7be83cdeb850b46d5cecccd3d70c94e"} Nov 29 00:09:23 crc kubenswrapper[4931]: I1129 00:09:23.734709 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:09:23 crc kubenswrapper[4931]: I1129 00:09:23.735187 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:09:53 crc kubenswrapper[4931]: I1129 00:09:53.735959 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:09:53 crc kubenswrapper[4931]: I1129 00:09:53.736683 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:10:23 crc kubenswrapper[4931]: I1129 00:10:23.734983 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:10:23 crc kubenswrapper[4931]: I1129 00:10:23.735597 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:10:23 crc kubenswrapper[4931]: I1129 00:10:23.735651 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:10:23 crc kubenswrapper[4931]: I1129 00:10:23.736422 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ffd259621969111fa5cf386e3c4f80aaf7be83cdeb850b46d5cecccd3d70c94e"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 00:10:23 crc kubenswrapper[4931]: I1129 00:10:23.736493 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://ffd259621969111fa5cf386e3c4f80aaf7be83cdeb850b46d5cecccd3d70c94e" gracePeriod=600 Nov 29 00:10:24 crc kubenswrapper[4931]: I1129 00:10:24.321638 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="ffd259621969111fa5cf386e3c4f80aaf7be83cdeb850b46d5cecccd3d70c94e" exitCode=0 Nov 29 00:10:24 crc kubenswrapper[4931]: I1129 00:10:24.321738 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"ffd259621969111fa5cf386e3c4f80aaf7be83cdeb850b46d5cecccd3d70c94e"} Nov 29 00:10:24 crc kubenswrapper[4931]: I1129 00:10:24.321990 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"b74c2da42faa59386db14c3bb51aa4674c697dd1c4383118b269448576e97f17"} Nov 29 00:10:24 crc kubenswrapper[4931]: I1129 00:10:24.322028 4931 scope.go:117] "RemoveContainer" containerID="720e0a368eab3734412726774de172ffe278950e85b711a795977fe783122261" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.480950 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-g6bkx"] Nov 29 00:11:11 crc kubenswrapper[4931]: E1129 00:11:11.481734 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93297fe9-2486-4b24-b6c6-cb917ab240c5" containerName="registry" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.481752 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="93297fe9-2486-4b24-b6c6-cb917ab240c5" containerName="registry" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.481882 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="93297fe9-2486-4b24-b6c6-cb917ab240c5" containerName="registry" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.482328 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-g6bkx" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.485045 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.485283 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.486976 4931 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-nrfzr" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.488668 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-s9fhh"] Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.489340 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-s9fhh" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.490867 4931 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-jzm5h" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.491223 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64fsg\" (UniqueName: \"kubernetes.io/projected/2f242144-c19c-4886-a51b-8547f684e2e9-kube-api-access-64fsg\") pod \"cert-manager-cainjector-7f985d654d-g6bkx\" (UID: \"2f242144-c19c-4886-a51b-8547f684e2e9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-g6bkx" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.508787 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-s9fhh"] Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.511506 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-b8cd4"] Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.512088 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-b8cd4" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.518318 4931 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-htp4c" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.522545 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-b8cd4"] Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.533421 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-g6bkx"] Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.592815 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64fsg\" (UniqueName: \"kubernetes.io/projected/2f242144-c19c-4886-a51b-8547f684e2e9-kube-api-access-64fsg\") pod \"cert-manager-cainjector-7f985d654d-g6bkx\" (UID: \"2f242144-c19c-4886-a51b-8547f684e2e9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-g6bkx" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.592895 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcb76\" (UniqueName: \"kubernetes.io/projected/283e1d51-cb76-418b-b4b0-d444320900ff-kube-api-access-xcb76\") pod \"cert-manager-webhook-5655c58dd6-b8cd4\" (UID: \"283e1d51-cb76-418b-b4b0-d444320900ff\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-b8cd4" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.592929 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls2nf\" (UniqueName: \"kubernetes.io/projected/ea00643b-3840-4675-b14a-323b377e2339-kube-api-access-ls2nf\") pod \"cert-manager-5b446d88c5-s9fhh\" (UID: \"ea00643b-3840-4675-b14a-323b377e2339\") " pod="cert-manager/cert-manager-5b446d88c5-s9fhh" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.614350 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64fsg\" (UniqueName: \"kubernetes.io/projected/2f242144-c19c-4886-a51b-8547f684e2e9-kube-api-access-64fsg\") pod \"cert-manager-cainjector-7f985d654d-g6bkx\" (UID: \"2f242144-c19c-4886-a51b-8547f684e2e9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-g6bkx" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.694095 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcb76\" (UniqueName: \"kubernetes.io/projected/283e1d51-cb76-418b-b4b0-d444320900ff-kube-api-access-xcb76\") pod \"cert-manager-webhook-5655c58dd6-b8cd4\" (UID: \"283e1d51-cb76-418b-b4b0-d444320900ff\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-b8cd4" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.694169 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls2nf\" (UniqueName: \"kubernetes.io/projected/ea00643b-3840-4675-b14a-323b377e2339-kube-api-access-ls2nf\") pod \"cert-manager-5b446d88c5-s9fhh\" (UID: \"ea00643b-3840-4675-b14a-323b377e2339\") " pod="cert-manager/cert-manager-5b446d88c5-s9fhh" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.711891 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcb76\" (UniqueName: \"kubernetes.io/projected/283e1d51-cb76-418b-b4b0-d444320900ff-kube-api-access-xcb76\") pod \"cert-manager-webhook-5655c58dd6-b8cd4\" (UID: \"283e1d51-cb76-418b-b4b0-d444320900ff\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-b8cd4" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.713765 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls2nf\" (UniqueName: \"kubernetes.io/projected/ea00643b-3840-4675-b14a-323b377e2339-kube-api-access-ls2nf\") pod \"cert-manager-5b446d88c5-s9fhh\" (UID: \"ea00643b-3840-4675-b14a-323b377e2339\") " pod="cert-manager/cert-manager-5b446d88c5-s9fhh" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.802268 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-g6bkx" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.808366 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-s9fhh" Nov 29 00:11:11 crc kubenswrapper[4931]: I1129 00:11:11.835462 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-b8cd4" Nov 29 00:11:12 crc kubenswrapper[4931]: I1129 00:11:12.222688 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-g6bkx"] Nov 29 00:11:12 crc kubenswrapper[4931]: I1129 00:11:12.228612 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 00:11:12 crc kubenswrapper[4931]: I1129 00:11:12.285340 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-s9fhh"] Nov 29 00:11:12 crc kubenswrapper[4931]: W1129 00:11:12.297757 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea00643b_3840_4675_b14a_323b377e2339.slice/crio-4f16d6f6c2e4db83f240f6293b03b1e72ff1582fac8a257055083de7e9d3f51b WatchSource:0}: Error finding container 4f16d6f6c2e4db83f240f6293b03b1e72ff1582fac8a257055083de7e9d3f51b: Status 404 returned error can't find the container with id 4f16d6f6c2e4db83f240f6293b03b1e72ff1582fac8a257055083de7e9d3f51b Nov 29 00:11:12 crc kubenswrapper[4931]: I1129 00:11:12.303884 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-b8cd4"] Nov 29 00:11:12 crc kubenswrapper[4931]: I1129 00:11:12.599401 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-s9fhh" event={"ID":"ea00643b-3840-4675-b14a-323b377e2339","Type":"ContainerStarted","Data":"4f16d6f6c2e4db83f240f6293b03b1e72ff1582fac8a257055083de7e9d3f51b"} Nov 29 00:11:12 crc kubenswrapper[4931]: I1129 00:11:12.600424 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-b8cd4" event={"ID":"283e1d51-cb76-418b-b4b0-d444320900ff","Type":"ContainerStarted","Data":"bc7a4752e02e2efea1fe8577f738d461922252e8e8f0bd356994c6425d405fb7"} Nov 29 00:11:12 crc kubenswrapper[4931]: I1129 00:11:12.601575 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-g6bkx" event={"ID":"2f242144-c19c-4886-a51b-8547f684e2e9","Type":"ContainerStarted","Data":"19151b11884b07b4ec1066f34eb665a9a4ae851789795beaf4a2d4f9528a687d"} Nov 29 00:11:25 crc kubenswrapper[4931]: I1129 00:11:25.688595 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-g6bkx" event={"ID":"2f242144-c19c-4886-a51b-8547f684e2e9","Type":"ContainerStarted","Data":"2177d66f4c02757959c5a730f9800bcc03309189b6f2244026d1d88529e88675"} Nov 29 00:11:25 crc kubenswrapper[4931]: I1129 00:11:25.690359 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-s9fhh" event={"ID":"ea00643b-3840-4675-b14a-323b377e2339","Type":"ContainerStarted","Data":"fa2a56fa53ef75092274e5dc377eb673888c310c2c3c82b29a8fb0a61fbe7f4d"} Nov 29 00:11:25 crc kubenswrapper[4931]: I1129 00:11:25.691586 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-b8cd4" event={"ID":"283e1d51-cb76-418b-b4b0-d444320900ff","Type":"ContainerStarted","Data":"5252ce889d70f2741eee1fdce0290de3dc8be3a57c8d6398fc6a27e4c9104096"} Nov 29 00:11:25 crc kubenswrapper[4931]: I1129 00:11:25.691748 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-b8cd4" Nov 29 00:11:25 crc kubenswrapper[4931]: I1129 00:11:25.700668 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-g6bkx" podStartSLOduration=2.278873699 podStartE2EDuration="14.700652038s" podCreationTimestamp="2025-11-29 00:11:11 +0000 UTC" firstStartedPulling="2025-11-29 00:11:12.228344148 +0000 UTC m=+685.390237370" lastFinishedPulling="2025-11-29 00:11:24.650122477 +0000 UTC m=+697.812015709" observedRunningTime="2025-11-29 00:11:25.700373519 +0000 UTC m=+698.862266771" watchObservedRunningTime="2025-11-29 00:11:25.700652038 +0000 UTC m=+698.862545270" Nov 29 00:11:25 crc kubenswrapper[4931]: I1129 00:11:25.716494 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-s9fhh" podStartSLOduration=2.311717384 podStartE2EDuration="14.716465867s" podCreationTimestamp="2025-11-29 00:11:11 +0000 UTC" firstStartedPulling="2025-11-29 00:11:12.301340189 +0000 UTC m=+685.463233431" lastFinishedPulling="2025-11-29 00:11:24.706088682 +0000 UTC m=+697.867981914" observedRunningTime="2025-11-29 00:11:25.713984772 +0000 UTC m=+698.875878014" watchObservedRunningTime="2025-11-29 00:11:25.716465867 +0000 UTC m=+698.878359109" Nov 29 00:11:25 crc kubenswrapper[4931]: I1129 00:11:25.732497 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-b8cd4" podStartSLOduration=2.3726877809999998 podStartE2EDuration="14.732477242s" podCreationTimestamp="2025-11-29 00:11:11 +0000 UTC" firstStartedPulling="2025-11-29 00:11:12.306008561 +0000 UTC m=+685.467901783" lastFinishedPulling="2025-11-29 00:11:24.665798012 +0000 UTC m=+697.827691244" observedRunningTime="2025-11-29 00:11:25.73242963 +0000 UTC m=+698.894322902" watchObservedRunningTime="2025-11-29 00:11:25.732477242 +0000 UTC m=+698.894370474" Nov 29 00:11:31 crc kubenswrapper[4931]: I1129 00:11:31.918263 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-b8cd4" Nov 29 00:11:44 crc kubenswrapper[4931]: I1129 00:11:44.382307 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4m4l4"] Nov 29 00:11:44 crc kubenswrapper[4931]: I1129 00:11:44.383198 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovn-controller" containerID="cri-o://ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91" gracePeriod=30 Nov 29 00:11:44 crc kubenswrapper[4931]: I1129 00:11:44.383539 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="sbdb" containerID="cri-o://ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0" gracePeriod=30 Nov 29 00:11:44 crc kubenswrapper[4931]: I1129 00:11:44.383575 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="nbdb" containerID="cri-o://4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243" gracePeriod=30 Nov 29 00:11:44 crc kubenswrapper[4931]: I1129 00:11:44.383605 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="northd" containerID="cri-o://6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139" gracePeriod=30 Nov 29 00:11:44 crc kubenswrapper[4931]: I1129 00:11:44.383658 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879" gracePeriod=30 Nov 29 00:11:44 crc kubenswrapper[4931]: I1129 00:11:44.383690 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="kube-rbac-proxy-node" containerID="cri-o://0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93" gracePeriod=30 Nov 29 00:11:44 crc kubenswrapper[4931]: I1129 00:11:44.383716 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovn-acl-logging" containerID="cri-o://068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40" gracePeriod=30 Nov 29 00:11:44 crc kubenswrapper[4931]: I1129 00:11:44.416364 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" containerID="cri-o://23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946" gracePeriod=30 Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.151925 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/3.log" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.154222 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovn-acl-logging/0.log" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.154637 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovn-controller/0.log" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.155040 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.226526 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hqtpq"] Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.226846 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="sbdb" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.226864 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="sbdb" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.226882 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.226890 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.226900 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovn-acl-logging" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.226907 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovn-acl-logging" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.226920 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="kubecfg-setup" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.226978 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="kubecfg-setup" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.226995 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="nbdb" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227005 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="nbdb" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.227014 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227023 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.227033 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227040 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.227048 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovn-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227054 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovn-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.227065 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227071 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.227080 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="northd" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227087 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="northd" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.227110 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="kube-rbac-proxy-ovn-metrics" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227118 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="kube-rbac-proxy-ovn-metrics" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.227129 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="kube-rbac-proxy-node" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227136 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="kube-rbac-proxy-node" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227269 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227282 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovn-acl-logging" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227293 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227302 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovn-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227311 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="kube-rbac-proxy-ovn-metrics" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227318 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227328 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227336 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="northd" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227349 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="nbdb" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227360 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="kube-rbac-proxy-node" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227369 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="sbdb" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.227525 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227536 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.227650 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerName="ovnkube-controller" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.229732 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334477 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-systemd\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334524 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfsrg\" (UniqueName: \"kubernetes.io/projected/2fae9705-fac9-4c99-b5f8-280e5c05e450-kube-api-access-nfsrg\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334555 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-openvswitch\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334577 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-cni-bin\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334602 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovn-node-metrics-cert\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334621 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-etc-openvswitch\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334660 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-run-netns\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334683 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-var-lib-cni-networks-ovn-kubernetes\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334702 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-cni-netd\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334723 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-ovn\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334743 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-run-ovn-kubernetes\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334781 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-env-overrides\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334805 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-kubelet\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334855 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-var-lib-openvswitch\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334878 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovnkube-config\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334902 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-slash\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334919 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-node-log\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334935 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-systemd-units\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334956 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovnkube-script-lib\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334975 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-log-socket\") pod \"2fae9705-fac9-4c99-b5f8-280e5c05e450\" (UID: \"2fae9705-fac9-4c99-b5f8-280e5c05e450\") " Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334674 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335024 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334717 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.334747 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335100 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335140 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-slash" (OuterVolumeSpecName: "host-slash") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335153 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335147 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335168 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-log-socket" (OuterVolumeSpecName: "log-socket") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335173 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335159 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335203 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335234 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-node-log" (OuterVolumeSpecName: "node-log") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335260 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335403 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335477 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335504 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335603 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-log-socket\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335667 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-run-openvswitch\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335710 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-node-log\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.335931 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-ovnkube-script-lib\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336105 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-run-ovn\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336148 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-run-ovn-kubernetes\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336175 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8lkt\" (UniqueName: \"kubernetes.io/projected/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-kube-api-access-n8lkt\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336235 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-kubelet\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336302 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-var-lib-openvswitch\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336384 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-env-overrides\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336452 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-systemd-units\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336518 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-run-netns\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336583 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-run-systemd\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336657 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-ovnkube-config\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336697 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-cni-bin\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336731 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-etc-openvswitch\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336766 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-cni-netd\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336801 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336862 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-ovn-node-metrics-cert\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336907 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-slash\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.336999 4931 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337021 4931 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337040 4931 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337057 4931 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337073 4931 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-node-log\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337089 4931 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-slash\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337109 4931 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337151 4931 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337175 4931 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-log-socket\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337202 4931 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337224 4931 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337248 4931 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337269 4931 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337297 4931 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337323 4931 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337347 4931 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.337369 4931 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.343299 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.343350 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fae9705-fac9-4c99-b5f8-280e5c05e450-kube-api-access-nfsrg" (OuterVolumeSpecName: "kube-api-access-nfsrg") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "kube-api-access-nfsrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.347913 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "2fae9705-fac9-4c99-b5f8-280e5c05e450" (UID: "2fae9705-fac9-4c99-b5f8-280e5c05e450"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438618 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8lkt\" (UniqueName: \"kubernetes.io/projected/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-kube-api-access-n8lkt\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438691 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-kubelet\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438719 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-var-lib-openvswitch\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438740 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-env-overrides\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438761 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-systemd-units\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438787 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-run-netns\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438827 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-run-systemd\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438852 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-ovnkube-config\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438878 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-cni-bin\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438897 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-etc-openvswitch\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438921 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-cni-netd\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438945 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438968 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-ovn-node-metrics-cert\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.438990 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-slash\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439015 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-log-socket\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439035 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-run-openvswitch\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439062 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-node-log\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439084 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-ovnkube-script-lib\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439126 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-run-ovn\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439147 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-run-ovn-kubernetes\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439190 4931 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2fae9705-fac9-4c99-b5f8-280e5c05e450-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439203 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfsrg\" (UniqueName: \"kubernetes.io/projected/2fae9705-fac9-4c99-b5f8-280e5c05e450-kube-api-access-nfsrg\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439215 4931 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2fae9705-fac9-4c99-b5f8-280e5c05e450-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439269 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-run-ovn-kubernetes\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439618 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-cni-netd\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439644 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-etc-openvswitch\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439657 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439752 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-kubelet\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.439786 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-var-lib-openvswitch\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.440198 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-run-systemd\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.440237 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-systemd-units\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.440261 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-run-netns\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.440520 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-env-overrides\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.440591 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-cni-bin\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.440642 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-run-openvswitch\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.440697 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-host-slash\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.440728 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-ovnkube-config\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.440730 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-log-socket\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.440769 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-node-log\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.440801 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-run-ovn\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.441188 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-ovnkube-script-lib\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.444354 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-ovn-node-metrics-cert\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.463033 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8lkt\" (UniqueName: \"kubernetes.io/projected/381fdf9a-e6d9-4391-bb14-35892b8ef2ce-kube-api-access-n8lkt\") pod \"ovnkube-node-hqtpq\" (UID: \"381fdf9a-e6d9-4391-bb14-35892b8ef2ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.548912 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:45 crc kubenswrapper[4931]: W1129 00:11:45.578001 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod381fdf9a_e6d9_4391_bb14_35892b8ef2ce.slice/crio-bddb8c6a0b530aef825c26ff6d7afc1dbee70827436826e060cafee3cfa523af WatchSource:0}: Error finding container bddb8c6a0b530aef825c26ff6d7afc1dbee70827436826e060cafee3cfa523af: Status 404 returned error can't find the container with id bddb8c6a0b530aef825c26ff6d7afc1dbee70827436826e060cafee3cfa523af Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.804075 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovnkube-controller/3.log" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.806527 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovn-acl-logging/0.log" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.806966 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4m4l4_2fae9705-fac9-4c99-b5f8-280e5c05e450/ovn-controller/0.log" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807309 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946" exitCode=0 Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807346 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0" exitCode=0 Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807356 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243" exitCode=0 Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807363 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139" exitCode=0 Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807370 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879" exitCode=0 Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807378 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93" exitCode=0 Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807385 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40" exitCode=143 Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807401 4931 generic.go:334] "Generic (PLEG): container finished" podID="2fae9705-fac9-4c99-b5f8-280e5c05e450" containerID="ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91" exitCode=143 Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807410 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807396 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807471 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807486 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807498 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807510 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807520 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807531 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807541 4931 scope.go:117] "RemoveContainer" containerID="23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807544 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807642 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807654 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807661 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807668 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807674 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807682 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807687 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807702 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807714 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807720 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807726 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807732 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807737 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807741 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807746 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807751 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807757 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807762 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807768 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807791 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807799 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807821 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807828 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807835 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807841 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807847 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807853 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807859 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807865 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807875 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4m4l4" event={"ID":"2fae9705-fac9-4c99-b5f8-280e5c05e450","Type":"ContainerDied","Data":"1e64dbe7efef590a70b3fa79084f97cb036c5a7dbd9bca3bebb5759f774a2662"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807887 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807894 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807901 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807908 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807914 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807919 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807925 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807931 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807937 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.807943 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.809519 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22vkg_31e6835b-462a-4a4f-9ba9-74531c14d41a/kube-multus/2.log" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.810043 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22vkg_31e6835b-462a-4a4f-9ba9-74531c14d41a/kube-multus/1.log" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.810084 4931 generic.go:334] "Generic (PLEG): container finished" podID="31e6835b-462a-4a4f-9ba9-74531c14d41a" containerID="8759f244d50ee7187a9b3f8c435be8a5a62e4a10fb675c1dc1e1a29cac90118b" exitCode=2 Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.810247 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22vkg" event={"ID":"31e6835b-462a-4a4f-9ba9-74531c14d41a","Type":"ContainerDied","Data":"8759f244d50ee7187a9b3f8c435be8a5a62e4a10fb675c1dc1e1a29cac90118b"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.810274 4931 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.811688 4931 generic.go:334] "Generic (PLEG): container finished" podID="381fdf9a-e6d9-4391-bb14-35892b8ef2ce" containerID="21653529c43a88dd0cedaf643e4a8782fed91b33d8cf40c2d0472501312f4055" exitCode=0 Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.811723 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" event={"ID":"381fdf9a-e6d9-4391-bb14-35892b8ef2ce","Type":"ContainerDied","Data":"21653529c43a88dd0cedaf643e4a8782fed91b33d8cf40c2d0472501312f4055"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.811755 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" event={"ID":"381fdf9a-e6d9-4391-bb14-35892b8ef2ce","Type":"ContainerStarted","Data":"bddb8c6a0b530aef825c26ff6d7afc1dbee70827436826e060cafee3cfa523af"} Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.813564 4931 scope.go:117] "RemoveContainer" containerID="8759f244d50ee7187a9b3f8c435be8a5a62e4a10fb675c1dc1e1a29cac90118b" Nov 29 00:11:45 crc kubenswrapper[4931]: E1129 00:11:45.814198 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-22vkg_openshift-multus(31e6835b-462a-4a4f-9ba9-74531c14d41a)\"" pod="openshift-multus/multus-22vkg" podUID="31e6835b-462a-4a4f-9ba9-74531c14d41a" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.828206 4931 scope.go:117] "RemoveContainer" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.846327 4931 scope.go:117] "RemoveContainer" containerID="ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.869437 4931 scope.go:117] "RemoveContainer" containerID="4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.883525 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4m4l4"] Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.892984 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4m4l4"] Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.898183 4931 scope.go:117] "RemoveContainer" containerID="6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.914948 4931 scope.go:117] "RemoveContainer" containerID="56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.935477 4931 scope.go:117] "RemoveContainer" containerID="0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.960048 4931 scope.go:117] "RemoveContainer" containerID="068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.979794 4931 scope.go:117] "RemoveContainer" containerID="ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91" Nov 29 00:11:45 crc kubenswrapper[4931]: I1129 00:11:45.995847 4931 scope.go:117] "RemoveContainer" containerID="b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.008563 4931 scope.go:117] "RemoveContainer" containerID="23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946" Nov 29 00:11:46 crc kubenswrapper[4931]: E1129 00:11:46.008880 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946\": container with ID starting with 23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946 not found: ID does not exist" containerID="23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.008907 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946"} err="failed to get container status \"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946\": rpc error: code = NotFound desc = could not find container \"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946\": container with ID starting with 23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.008927 4931 scope.go:117] "RemoveContainer" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" Nov 29 00:11:46 crc kubenswrapper[4931]: E1129 00:11:46.009211 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\": container with ID starting with 772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c not found: ID does not exist" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.009245 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c"} err="failed to get container status \"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\": rpc error: code = NotFound desc = could not find container \"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\": container with ID starting with 772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.009269 4931 scope.go:117] "RemoveContainer" containerID="ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0" Nov 29 00:11:46 crc kubenswrapper[4931]: E1129 00:11:46.009556 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\": container with ID starting with ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0 not found: ID does not exist" containerID="ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.009578 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0"} err="failed to get container status \"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\": rpc error: code = NotFound desc = could not find container \"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\": container with ID starting with ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.009592 4931 scope.go:117] "RemoveContainer" containerID="4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243" Nov 29 00:11:46 crc kubenswrapper[4931]: E1129 00:11:46.009857 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\": container with ID starting with 4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243 not found: ID does not exist" containerID="4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.009889 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243"} err="failed to get container status \"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\": rpc error: code = NotFound desc = could not find container \"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\": container with ID starting with 4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.009910 4931 scope.go:117] "RemoveContainer" containerID="6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139" Nov 29 00:11:46 crc kubenswrapper[4931]: E1129 00:11:46.010213 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\": container with ID starting with 6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139 not found: ID does not exist" containerID="6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.010233 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139"} err="failed to get container status \"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\": rpc error: code = NotFound desc = could not find container \"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\": container with ID starting with 6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.010246 4931 scope.go:117] "RemoveContainer" containerID="56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879" Nov 29 00:11:46 crc kubenswrapper[4931]: E1129 00:11:46.010447 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\": container with ID starting with 56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879 not found: ID does not exist" containerID="56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.010463 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879"} err="failed to get container status \"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\": rpc error: code = NotFound desc = could not find container \"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\": container with ID starting with 56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.010474 4931 scope.go:117] "RemoveContainer" containerID="0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93" Nov 29 00:11:46 crc kubenswrapper[4931]: E1129 00:11:46.010652 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\": container with ID starting with 0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93 not found: ID does not exist" containerID="0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.010670 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93"} err="failed to get container status \"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\": rpc error: code = NotFound desc = could not find container \"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\": container with ID starting with 0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.010684 4931 scope.go:117] "RemoveContainer" containerID="068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40" Nov 29 00:11:46 crc kubenswrapper[4931]: E1129 00:11:46.010904 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\": container with ID starting with 068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40 not found: ID does not exist" containerID="068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.010927 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40"} err="failed to get container status \"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\": rpc error: code = NotFound desc = could not find container \"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\": container with ID starting with 068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.010939 4931 scope.go:117] "RemoveContainer" containerID="ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91" Nov 29 00:11:46 crc kubenswrapper[4931]: E1129 00:11:46.012594 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\": container with ID starting with ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91 not found: ID does not exist" containerID="ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.012617 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91"} err="failed to get container status \"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\": rpc error: code = NotFound desc = could not find container \"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\": container with ID starting with ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.012629 4931 scope.go:117] "RemoveContainer" containerID="b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e" Nov 29 00:11:46 crc kubenswrapper[4931]: E1129 00:11:46.013051 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\": container with ID starting with b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e not found: ID does not exist" containerID="b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.013075 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e"} err="failed to get container status \"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\": rpc error: code = NotFound desc = could not find container \"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\": container with ID starting with b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.013090 4931 scope.go:117] "RemoveContainer" containerID="23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.013358 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946"} err="failed to get container status \"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946\": rpc error: code = NotFound desc = could not find container \"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946\": container with ID starting with 23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.013380 4931 scope.go:117] "RemoveContainer" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.013710 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c"} err="failed to get container status \"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\": rpc error: code = NotFound desc = could not find container \"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\": container with ID starting with 772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.013743 4931 scope.go:117] "RemoveContainer" containerID="ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.013981 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0"} err="failed to get container status \"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\": rpc error: code = NotFound desc = could not find container \"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\": container with ID starting with ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.014058 4931 scope.go:117] "RemoveContainer" containerID="4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.014284 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243"} err="failed to get container status \"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\": rpc error: code = NotFound desc = could not find container \"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\": container with ID starting with 4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.014304 4931 scope.go:117] "RemoveContainer" containerID="6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.014469 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139"} err="failed to get container status \"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\": rpc error: code = NotFound desc = could not find container \"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\": container with ID starting with 6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.014491 4931 scope.go:117] "RemoveContainer" containerID="56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.014781 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879"} err="failed to get container status \"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\": rpc error: code = NotFound desc = could not find container \"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\": container with ID starting with 56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.014797 4931 scope.go:117] "RemoveContainer" containerID="0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.015043 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93"} err="failed to get container status \"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\": rpc error: code = NotFound desc = could not find container \"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\": container with ID starting with 0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.015055 4931 scope.go:117] "RemoveContainer" containerID="068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.015194 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40"} err="failed to get container status \"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\": rpc error: code = NotFound desc = could not find container \"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\": container with ID starting with 068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.015206 4931 scope.go:117] "RemoveContainer" containerID="ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.015327 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91"} err="failed to get container status \"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\": rpc error: code = NotFound desc = could not find container \"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\": container with ID starting with ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.015340 4931 scope.go:117] "RemoveContainer" containerID="b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.015456 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e"} err="failed to get container status \"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\": rpc error: code = NotFound desc = could not find container \"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\": container with ID starting with b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.015471 4931 scope.go:117] "RemoveContainer" containerID="23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.027766 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946"} err="failed to get container status \"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946\": rpc error: code = NotFound desc = could not find container \"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946\": container with ID starting with 23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.027825 4931 scope.go:117] "RemoveContainer" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.028458 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c"} err="failed to get container status \"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\": rpc error: code = NotFound desc = could not find container \"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\": container with ID starting with 772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.028479 4931 scope.go:117] "RemoveContainer" containerID="ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.028701 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0"} err="failed to get container status \"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\": rpc error: code = NotFound desc = could not find container \"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\": container with ID starting with ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.028740 4931 scope.go:117] "RemoveContainer" containerID="4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.028987 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243"} err="failed to get container status \"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\": rpc error: code = NotFound desc = could not find container \"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\": container with ID starting with 4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.029007 4931 scope.go:117] "RemoveContainer" containerID="6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.029215 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139"} err="failed to get container status \"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\": rpc error: code = NotFound desc = could not find container \"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\": container with ID starting with 6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.029235 4931 scope.go:117] "RemoveContainer" containerID="56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.029535 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879"} err="failed to get container status \"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\": rpc error: code = NotFound desc = could not find container \"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\": container with ID starting with 56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.029552 4931 scope.go:117] "RemoveContainer" containerID="0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.029756 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93"} err="failed to get container status \"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\": rpc error: code = NotFound desc = could not find container \"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\": container with ID starting with 0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.029773 4931 scope.go:117] "RemoveContainer" containerID="068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.030112 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40"} err="failed to get container status \"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\": rpc error: code = NotFound desc = could not find container \"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\": container with ID starting with 068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.030132 4931 scope.go:117] "RemoveContainer" containerID="ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.030378 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91"} err="failed to get container status \"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\": rpc error: code = NotFound desc = could not find container \"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\": container with ID starting with ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.030394 4931 scope.go:117] "RemoveContainer" containerID="b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.030706 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e"} err="failed to get container status \"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\": rpc error: code = NotFound desc = could not find container \"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\": container with ID starting with b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.030731 4931 scope.go:117] "RemoveContainer" containerID="23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.030998 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946"} err="failed to get container status \"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946\": rpc error: code = NotFound desc = could not find container \"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946\": container with ID starting with 23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.031015 4931 scope.go:117] "RemoveContainer" containerID="772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.031339 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c"} err="failed to get container status \"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\": rpc error: code = NotFound desc = could not find container \"772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c\": container with ID starting with 772a4715efdf2356de391eb3a60440e04b1181588a23adb6895e2dde21ac2b6c not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.031398 4931 scope.go:117] "RemoveContainer" containerID="ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.031781 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0"} err="failed to get container status \"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\": rpc error: code = NotFound desc = could not find container \"ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0\": container with ID starting with ab5320980961f21834dd6090ab6f04e0e9ec0a2df49cf131781326045ba41cd0 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.031800 4931 scope.go:117] "RemoveContainer" containerID="4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.032057 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243"} err="failed to get container status \"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\": rpc error: code = NotFound desc = could not find container \"4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243\": container with ID starting with 4ff617de17f75cd7f75a4a2e20afcf7b65b481d0202f5b9b2fb4414a4326b243 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.032080 4931 scope.go:117] "RemoveContainer" containerID="6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.032340 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139"} err="failed to get container status \"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\": rpc error: code = NotFound desc = could not find container \"6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139\": container with ID starting with 6f563f50ade856a426fa3e4db91ab01dbb7bc2a984e874d11962148d4c0f6139 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.032357 4931 scope.go:117] "RemoveContainer" containerID="56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.032617 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879"} err="failed to get container status \"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\": rpc error: code = NotFound desc = could not find container \"56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879\": container with ID starting with 56667a995715b29f2df831726052959774554a943252260dfb2e9f7f9df22879 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.032634 4931 scope.go:117] "RemoveContainer" containerID="0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.032869 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93"} err="failed to get container status \"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\": rpc error: code = NotFound desc = could not find container \"0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93\": container with ID starting with 0ca6a71f3085c8086d580056c44bb5256a0d1f41a013da81c19cadd955319e93 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.032884 4931 scope.go:117] "RemoveContainer" containerID="068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.033123 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40"} err="failed to get container status \"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\": rpc error: code = NotFound desc = could not find container \"068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40\": container with ID starting with 068d701e5c86c231b69cff87efcbd1961e78d1a06d554cac9e810ccc1df20d40 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.033139 4931 scope.go:117] "RemoveContainer" containerID="ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.033341 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91"} err="failed to get container status \"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\": rpc error: code = NotFound desc = could not find container \"ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91\": container with ID starting with ac787a5b6f4e8df83f727d490fc5f4db0239becbdc04ea34ea725b4daeccff91 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.033357 4931 scope.go:117] "RemoveContainer" containerID="b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.033558 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e"} err="failed to get container status \"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\": rpc error: code = NotFound desc = could not find container \"b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e\": container with ID starting with b6904194bbe3dd9b54738f5a7c36561319b11650be6e03a460656dd2ac8ae01e not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.033574 4931 scope.go:117] "RemoveContainer" containerID="23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.033850 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946"} err="failed to get container status \"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946\": rpc error: code = NotFound desc = could not find container \"23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946\": container with ID starting with 23badb8886228ae02866347ab2b3e50ffe194483fc0b2c75648c7ab083175946 not found: ID does not exist" Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.825366 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" event={"ID":"381fdf9a-e6d9-4391-bb14-35892b8ef2ce","Type":"ContainerStarted","Data":"8292c221ee20cd6d70675243d68d6ea3dc8f04dbcef1fe067a1a5689f15ba496"} Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.825837 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" event={"ID":"381fdf9a-e6d9-4391-bb14-35892b8ef2ce","Type":"ContainerStarted","Data":"3f55ad1211b6125fa75e23d910099209ebb583b94d98b6a96a870c5489d76439"} Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.825860 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" event={"ID":"381fdf9a-e6d9-4391-bb14-35892b8ef2ce","Type":"ContainerStarted","Data":"463b88cc1eb68023d95e67a55efef586d761d7eb5440c4340bd612c2195bf9a9"} Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.825877 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" event={"ID":"381fdf9a-e6d9-4391-bb14-35892b8ef2ce","Type":"ContainerStarted","Data":"bf0f0cf820a34f541ddcd5fdd400bdfe908a842c6e916788428a0c2e61e5b5ea"} Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.825889 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" event={"ID":"381fdf9a-e6d9-4391-bb14-35892b8ef2ce","Type":"ContainerStarted","Data":"8426581366fcbc1a6b31e3548eb1248ffdf69946c4d4df892884889d0068c0f7"} Nov 29 00:11:46 crc kubenswrapper[4931]: I1129 00:11:46.825901 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" event={"ID":"381fdf9a-e6d9-4391-bb14-35892b8ef2ce","Type":"ContainerStarted","Data":"c09a6850ccdee7d2ac02889e54ffcbeca4e5e9e5c335449929ff308a7689c15c"} Nov 29 00:11:47 crc kubenswrapper[4931]: I1129 00:11:47.226570 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fae9705-fac9-4c99-b5f8-280e5c05e450" path="/var/lib/kubelet/pods/2fae9705-fac9-4c99-b5f8-280e5c05e450/volumes" Nov 29 00:11:47 crc kubenswrapper[4931]: I1129 00:11:47.507411 4931 scope.go:117] "RemoveContainer" containerID="7886c26a635a99e232c6ca5793eb0efd698520b387c43d01f10176b1a134c652" Nov 29 00:11:47 crc kubenswrapper[4931]: I1129 00:11:47.835368 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22vkg_31e6835b-462a-4a4f-9ba9-74531c14d41a/kube-multus/2.log" Nov 29 00:11:49 crc kubenswrapper[4931]: I1129 00:11:49.853145 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" event={"ID":"381fdf9a-e6d9-4391-bb14-35892b8ef2ce","Type":"ContainerStarted","Data":"d03ade2dffda33492177d32b53ee72e74779d9ed0a8c7131e374694460c6508f"} Nov 29 00:11:51 crc kubenswrapper[4931]: I1129 00:11:51.872596 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" event={"ID":"381fdf9a-e6d9-4391-bb14-35892b8ef2ce","Type":"ContainerStarted","Data":"f211741e7ca446b8df2dc929adc1fe10720f58b38df13f6569e87d9364e672d7"} Nov 29 00:11:51 crc kubenswrapper[4931]: I1129 00:11:51.873062 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:51 crc kubenswrapper[4931]: I1129 00:11:51.905947 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:51 crc kubenswrapper[4931]: I1129 00:11:51.924064 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" podStartSLOduration=6.924041364 podStartE2EDuration="6.924041364s" podCreationTimestamp="2025-11-29 00:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:11:51.916732248 +0000 UTC m=+725.078625490" watchObservedRunningTime="2025-11-29 00:11:51.924041364 +0000 UTC m=+725.085934626" Nov 29 00:11:52 crc kubenswrapper[4931]: I1129 00:11:52.881467 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:52 crc kubenswrapper[4931]: I1129 00:11:52.881524 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:52 crc kubenswrapper[4931]: I1129 00:11:52.956563 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:11:58 crc kubenswrapper[4931]: I1129 00:11:58.212454 4931 scope.go:117] "RemoveContainer" containerID="8759f244d50ee7187a9b3f8c435be8a5a62e4a10fb675c1dc1e1a29cac90118b" Nov 29 00:11:58 crc kubenswrapper[4931]: E1129 00:11:58.214192 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-22vkg_openshift-multus(31e6835b-462a-4a4f-9ba9-74531c14d41a)\"" pod="openshift-multus/multus-22vkg" podUID="31e6835b-462a-4a4f-9ba9-74531c14d41a" Nov 29 00:12:11 crc kubenswrapper[4931]: I1129 00:12:11.880392 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg"] Nov 29 00:12:11 crc kubenswrapper[4931]: I1129 00:12:11.882173 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:11 crc kubenswrapper[4931]: I1129 00:12:11.885075 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 29 00:12:11 crc kubenswrapper[4931]: I1129 00:12:11.893350 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg"] Nov 29 00:12:12 crc kubenswrapper[4931]: I1129 00:12:12.035037 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg\" (UID: \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:12 crc kubenswrapper[4931]: I1129 00:12:12.035689 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnx4g\" (UniqueName: \"kubernetes.io/projected/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-kube-api-access-wnx4g\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg\" (UID: \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:12 crc kubenswrapper[4931]: I1129 00:12:12.035786 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg\" (UID: \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:12 crc kubenswrapper[4931]: I1129 00:12:12.136954 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg\" (UID: \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:12 crc kubenswrapper[4931]: I1129 00:12:12.137435 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg\" (UID: \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:12 crc kubenswrapper[4931]: I1129 00:12:12.137369 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg\" (UID: \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:12 crc kubenswrapper[4931]: I1129 00:12:12.137705 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg\" (UID: \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:12 crc kubenswrapper[4931]: I1129 00:12:12.137881 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnx4g\" (UniqueName: \"kubernetes.io/projected/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-kube-api-access-wnx4g\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg\" (UID: \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:12 crc kubenswrapper[4931]: I1129 00:12:12.157285 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnx4g\" (UniqueName: \"kubernetes.io/projected/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-kube-api-access-wnx4g\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg\" (UID: \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:12 crc kubenswrapper[4931]: I1129 00:12:12.197046 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:12 crc kubenswrapper[4931]: E1129 00:12:12.219079 4931 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_openshift-marketplace_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00_0(3b45f28741be336e301b3c4cd77da26295b9a1d41d70fbcbe653a77104c2a82b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 29 00:12:12 crc kubenswrapper[4931]: E1129 00:12:12.219147 4931 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_openshift-marketplace_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00_0(3b45f28741be336e301b3c4cd77da26295b9a1d41d70fbcbe653a77104c2a82b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:12 crc kubenswrapper[4931]: E1129 00:12:12.219177 4931 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_openshift-marketplace_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00_0(3b45f28741be336e301b3c4cd77da26295b9a1d41d70fbcbe653a77104c2a82b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:12 crc kubenswrapper[4931]: E1129 00:12:12.219223 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_openshift-marketplace(990bc8d0-c4bf-4b25-acfb-87dcfde9dc00)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_openshift-marketplace(990bc8d0-c4bf-4b25-acfb-87dcfde9dc00)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_openshift-marketplace_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00_0(3b45f28741be336e301b3c4cd77da26295b9a1d41d70fbcbe653a77104c2a82b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" podUID="990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" Nov 29 00:12:13 crc kubenswrapper[4931]: I1129 00:12:13.000070 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:13 crc kubenswrapper[4931]: I1129 00:12:13.000915 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:13 crc kubenswrapper[4931]: E1129 00:12:13.027453 4931 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_openshift-marketplace_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00_0(34c2b25068c076f7a3ead3214b26d0f645a197562849e7254f28b4410458fe16): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 29 00:12:13 crc kubenswrapper[4931]: E1129 00:12:13.027530 4931 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_openshift-marketplace_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00_0(34c2b25068c076f7a3ead3214b26d0f645a197562849e7254f28b4410458fe16): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:13 crc kubenswrapper[4931]: E1129 00:12:13.027569 4931 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_openshift-marketplace_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00_0(34c2b25068c076f7a3ead3214b26d0f645a197562849e7254f28b4410458fe16): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:13 crc kubenswrapper[4931]: E1129 00:12:13.027635 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_openshift-marketplace(990bc8d0-c4bf-4b25-acfb-87dcfde9dc00)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_openshift-marketplace(990bc8d0-c4bf-4b25-acfb-87dcfde9dc00)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_openshift-marketplace_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00_0(34c2b25068c076f7a3ead3214b26d0f645a197562849e7254f28b4410458fe16): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" podUID="990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" Nov 29 00:12:13 crc kubenswrapper[4931]: I1129 00:12:13.212013 4931 scope.go:117] "RemoveContainer" containerID="8759f244d50ee7187a9b3f8c435be8a5a62e4a10fb675c1dc1e1a29cac90118b" Nov 29 00:12:14 crc kubenswrapper[4931]: I1129 00:12:14.010036 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-22vkg_31e6835b-462a-4a4f-9ba9-74531c14d41a/kube-multus/2.log" Nov 29 00:12:14 crc kubenswrapper[4931]: I1129 00:12:14.010132 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-22vkg" event={"ID":"31e6835b-462a-4a4f-9ba9-74531c14d41a","Type":"ContainerStarted","Data":"d230256a53cb9935cd2c9a6ace189a71ed316a22e33afbfb58861b1394c7bc87"} Nov 29 00:12:15 crc kubenswrapper[4931]: I1129 00:12:15.570326 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hqtpq" Nov 29 00:12:24 crc kubenswrapper[4931]: I1129 00:12:24.212348 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:24 crc kubenswrapper[4931]: I1129 00:12:24.214461 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:24 crc kubenswrapper[4931]: I1129 00:12:24.442915 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg"] Nov 29 00:12:25 crc kubenswrapper[4931]: I1129 00:12:25.071760 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" event={"ID":"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00","Type":"ContainerStarted","Data":"19407360e7bf7b5905cc4b657b11f4c73a07e9162e0206c96d1edf515cc05326"} Nov 29 00:12:26 crc kubenswrapper[4931]: I1129 00:12:26.079531 4931 generic.go:334] "Generic (PLEG): container finished" podID="990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" containerID="a8c64d8a3f05f94ad5bc244402f07185f92da6e8135e5d09afd178b5c7c09bfc" exitCode=0 Nov 29 00:12:26 crc kubenswrapper[4931]: I1129 00:12:26.079576 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" event={"ID":"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00","Type":"ContainerDied","Data":"a8c64d8a3f05f94ad5bc244402f07185f92da6e8135e5d09afd178b5c7c09bfc"} Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.093980 4931 generic.go:334] "Generic (PLEG): container finished" podID="990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" containerID="b4286085e6ee600c69f52ff953ea8978629a078707b19779aa07a50deaa3412c" exitCode=0 Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.094349 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" event={"ID":"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00","Type":"ContainerDied","Data":"b4286085e6ee600c69f52ff953ea8978629a078707b19779aa07a50deaa3412c"} Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.302073 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dkl9z"] Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.303070 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.308710 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dkl9z"] Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.453215 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1228beae-7460-4619-afe9-2835885b5bb9-utilities\") pod \"redhat-operators-dkl9z\" (UID: \"1228beae-7460-4619-afe9-2835885b5bb9\") " pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.453295 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1228beae-7460-4619-afe9-2835885b5bb9-catalog-content\") pod \"redhat-operators-dkl9z\" (UID: \"1228beae-7460-4619-afe9-2835885b5bb9\") " pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.453435 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-467n7\" (UniqueName: \"kubernetes.io/projected/1228beae-7460-4619-afe9-2835885b5bb9-kube-api-access-467n7\") pod \"redhat-operators-dkl9z\" (UID: \"1228beae-7460-4619-afe9-2835885b5bb9\") " pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.554367 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1228beae-7460-4619-afe9-2835885b5bb9-utilities\") pod \"redhat-operators-dkl9z\" (UID: \"1228beae-7460-4619-afe9-2835885b5bb9\") " pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.554415 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1228beae-7460-4619-afe9-2835885b5bb9-catalog-content\") pod \"redhat-operators-dkl9z\" (UID: \"1228beae-7460-4619-afe9-2835885b5bb9\") " pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.554473 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-467n7\" (UniqueName: \"kubernetes.io/projected/1228beae-7460-4619-afe9-2835885b5bb9-kube-api-access-467n7\") pod \"redhat-operators-dkl9z\" (UID: \"1228beae-7460-4619-afe9-2835885b5bb9\") " pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.555167 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1228beae-7460-4619-afe9-2835885b5bb9-utilities\") pod \"redhat-operators-dkl9z\" (UID: \"1228beae-7460-4619-afe9-2835885b5bb9\") " pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.555177 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1228beae-7460-4619-afe9-2835885b5bb9-catalog-content\") pod \"redhat-operators-dkl9z\" (UID: \"1228beae-7460-4619-afe9-2835885b5bb9\") " pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.574634 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-467n7\" (UniqueName: \"kubernetes.io/projected/1228beae-7460-4619-afe9-2835885b5bb9-kube-api-access-467n7\") pod \"redhat-operators-dkl9z\" (UID: \"1228beae-7460-4619-afe9-2835885b5bb9\") " pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:28 crc kubenswrapper[4931]: I1129 00:12:28.664926 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:29 crc kubenswrapper[4931]: I1129 00:12:29.073063 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dkl9z"] Nov 29 00:12:29 crc kubenswrapper[4931]: W1129 00:12:29.082191 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1228beae_7460_4619_afe9_2835885b5bb9.slice/crio-f9a4a152e3bb981addd803e67e81f793c1cbdcc66101acd218485cf97e38cba3 WatchSource:0}: Error finding container f9a4a152e3bb981addd803e67e81f793c1cbdcc66101acd218485cf97e38cba3: Status 404 returned error can't find the container with id f9a4a152e3bb981addd803e67e81f793c1cbdcc66101acd218485cf97e38cba3 Nov 29 00:12:29 crc kubenswrapper[4931]: I1129 00:12:29.098180 4931 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 29 00:12:29 crc kubenswrapper[4931]: I1129 00:12:29.100797 4931 generic.go:334] "Generic (PLEG): container finished" podID="990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" containerID="4fdc06e1d83a4d8ddd2a2c11fb3a32593a2cf8798effd8c56e52dff962a20a58" exitCode=0 Nov 29 00:12:29 crc kubenswrapper[4931]: I1129 00:12:29.100892 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" event={"ID":"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00","Type":"ContainerDied","Data":"4fdc06e1d83a4d8ddd2a2c11fb3a32593a2cf8798effd8c56e52dff962a20a58"} Nov 29 00:12:29 crc kubenswrapper[4931]: I1129 00:12:29.101722 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dkl9z" event={"ID":"1228beae-7460-4619-afe9-2835885b5bb9","Type":"ContainerStarted","Data":"f9a4a152e3bb981addd803e67e81f793c1cbdcc66101acd218485cf97e38cba3"} Nov 29 00:12:30 crc kubenswrapper[4931]: I1129 00:12:30.109121 4931 generic.go:334] "Generic (PLEG): container finished" podID="1228beae-7460-4619-afe9-2835885b5bb9" containerID="f9be5104476974554468c119b3ea875f5231c954e350ec243a0f7b4c50e25a30" exitCode=0 Nov 29 00:12:30 crc kubenswrapper[4931]: I1129 00:12:30.109212 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dkl9z" event={"ID":"1228beae-7460-4619-afe9-2835885b5bb9","Type":"ContainerDied","Data":"f9be5104476974554468c119b3ea875f5231c954e350ec243a0f7b4c50e25a30"} Nov 29 00:12:30 crc kubenswrapper[4931]: I1129 00:12:30.330916 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:30 crc kubenswrapper[4931]: I1129 00:12:30.477618 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnx4g\" (UniqueName: \"kubernetes.io/projected/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-kube-api-access-wnx4g\") pod \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\" (UID: \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\") " Nov 29 00:12:30 crc kubenswrapper[4931]: I1129 00:12:30.477689 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-util\") pod \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\" (UID: \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\") " Nov 29 00:12:30 crc kubenswrapper[4931]: I1129 00:12:30.477723 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-bundle\") pod \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\" (UID: \"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00\") " Nov 29 00:12:30 crc kubenswrapper[4931]: I1129 00:12:30.478387 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-bundle" (OuterVolumeSpecName: "bundle") pod "990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" (UID: "990bc8d0-c4bf-4b25-acfb-87dcfde9dc00"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:12:30 crc kubenswrapper[4931]: I1129 00:12:30.483239 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-kube-api-access-wnx4g" (OuterVolumeSpecName: "kube-api-access-wnx4g") pod "990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" (UID: "990bc8d0-c4bf-4b25-acfb-87dcfde9dc00"). InnerVolumeSpecName "kube-api-access-wnx4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:12:30 crc kubenswrapper[4931]: I1129 00:12:30.491179 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-util" (OuterVolumeSpecName: "util") pod "990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" (UID: "990bc8d0-c4bf-4b25-acfb-87dcfde9dc00"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:12:30 crc kubenswrapper[4931]: I1129 00:12:30.579454 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnx4g\" (UniqueName: \"kubernetes.io/projected/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-kube-api-access-wnx4g\") on node \"crc\" DevicePath \"\"" Nov 29 00:12:30 crc kubenswrapper[4931]: I1129 00:12:30.579492 4931 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-util\") on node \"crc\" DevicePath \"\"" Nov 29 00:12:30 crc kubenswrapper[4931]: I1129 00:12:30.579501 4931 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/990bc8d0-c4bf-4b25-acfb-87dcfde9dc00-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:12:31 crc kubenswrapper[4931]: I1129 00:12:31.115379 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" event={"ID":"990bc8d0-c4bf-4b25-acfb-87dcfde9dc00","Type":"ContainerDied","Data":"19407360e7bf7b5905cc4b657b11f4c73a07e9162e0206c96d1edf515cc05326"} Nov 29 00:12:31 crc kubenswrapper[4931]: I1129 00:12:31.115677 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19407360e7bf7b5905cc4b657b11f4c73a07e9162e0206c96d1edf515cc05326" Nov 29 00:12:31 crc kubenswrapper[4931]: I1129 00:12:31.115594 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg" Nov 29 00:12:31 crc kubenswrapper[4931]: I1129 00:12:31.117240 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dkl9z" event={"ID":"1228beae-7460-4619-afe9-2835885b5bb9","Type":"ContainerStarted","Data":"d62d0c6a4bdb0ab082f33a5d1fdd2364d63088b33f81e777788673066b0c7718"} Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.123870 4931 generic.go:334] "Generic (PLEG): container finished" podID="1228beae-7460-4619-afe9-2835885b5bb9" containerID="d62d0c6a4bdb0ab082f33a5d1fdd2364d63088b33f81e777788673066b0c7718" exitCode=0 Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.123956 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dkl9z" event={"ID":"1228beae-7460-4619-afe9-2835885b5bb9","Type":"ContainerDied","Data":"d62d0c6a4bdb0ab082f33a5d1fdd2364d63088b33f81e777788673066b0c7718"} Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.697101 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-t6vpl"] Nov 29 00:12:32 crc kubenswrapper[4931]: E1129 00:12:32.697294 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" containerName="util" Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.697307 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" containerName="util" Nov 29 00:12:32 crc kubenswrapper[4931]: E1129 00:12:32.697316 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" containerName="extract" Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.697321 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" containerName="extract" Nov 29 00:12:32 crc kubenswrapper[4931]: E1129 00:12:32.697333 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" containerName="pull" Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.697340 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" containerName="pull" Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.697424 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="990bc8d0-c4bf-4b25-acfb-87dcfde9dc00" containerName="extract" Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.697767 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t6vpl" Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.699452 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.699679 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-ldqsn" Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.702365 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.709501 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-t6vpl"] Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.804316 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmtv5\" (UniqueName: \"kubernetes.io/projected/3f3c7035-b1e3-48bf-bbb9-3c6400ffedae-kube-api-access-xmtv5\") pod \"nmstate-operator-5b5b58f5c8-t6vpl\" (UID: \"3f3c7035-b1e3-48bf-bbb9-3c6400ffedae\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t6vpl" Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.906002 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmtv5\" (UniqueName: \"kubernetes.io/projected/3f3c7035-b1e3-48bf-bbb9-3c6400ffedae-kube-api-access-xmtv5\") pod \"nmstate-operator-5b5b58f5c8-t6vpl\" (UID: \"3f3c7035-b1e3-48bf-bbb9-3c6400ffedae\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t6vpl" Nov 29 00:12:32 crc kubenswrapper[4931]: I1129 00:12:32.923053 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmtv5\" (UniqueName: \"kubernetes.io/projected/3f3c7035-b1e3-48bf-bbb9-3c6400ffedae-kube-api-access-xmtv5\") pod \"nmstate-operator-5b5b58f5c8-t6vpl\" (UID: \"3f3c7035-b1e3-48bf-bbb9-3c6400ffedae\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t6vpl" Nov 29 00:12:33 crc kubenswrapper[4931]: I1129 00:12:33.033066 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t6vpl" Nov 29 00:12:33 crc kubenswrapper[4931]: I1129 00:12:33.133262 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dkl9z" event={"ID":"1228beae-7460-4619-afe9-2835885b5bb9","Type":"ContainerStarted","Data":"9c651e6acd56615f9b7f0f43fbd7b1df28ccf7897bb659658d4d71f6f93304bf"} Nov 29 00:12:33 crc kubenswrapper[4931]: I1129 00:12:33.208065 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dkl9z" podStartSLOduration=2.682826694 podStartE2EDuration="5.208047556s" podCreationTimestamp="2025-11-29 00:12:28 +0000 UTC" firstStartedPulling="2025-11-29 00:12:30.112235855 +0000 UTC m=+763.274129117" lastFinishedPulling="2025-11-29 00:12:32.637456747 +0000 UTC m=+765.799349979" observedRunningTime="2025-11-29 00:12:33.150413081 +0000 UTC m=+766.312306323" watchObservedRunningTime="2025-11-29 00:12:33.208047556 +0000 UTC m=+766.369940808" Nov 29 00:12:33 crc kubenswrapper[4931]: I1129 00:12:33.219983 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-t6vpl"] Nov 29 00:12:34 crc kubenswrapper[4931]: I1129 00:12:34.140357 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t6vpl" event={"ID":"3f3c7035-b1e3-48bf-bbb9-3c6400ffedae","Type":"ContainerStarted","Data":"22419b478cdd2d05c09c82036d3584e6cda41162a57359b8fecb75111e4c10dc"} Nov 29 00:12:38 crc kubenswrapper[4931]: I1129 00:12:38.161337 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t6vpl" event={"ID":"3f3c7035-b1e3-48bf-bbb9-3c6400ffedae","Type":"ContainerStarted","Data":"8a76f07cce98dcab298bdb52b3fea318e780d01e51542d922763b06e077d31c0"} Nov 29 00:12:38 crc kubenswrapper[4931]: I1129 00:12:38.175720 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-t6vpl" podStartSLOduration=2.084722426 podStartE2EDuration="6.175704389s" podCreationTimestamp="2025-11-29 00:12:32 +0000 UTC" firstStartedPulling="2025-11-29 00:12:33.229174712 +0000 UTC m=+766.391067944" lastFinishedPulling="2025-11-29 00:12:37.320156665 +0000 UTC m=+770.482049907" observedRunningTime="2025-11-29 00:12:38.173330282 +0000 UTC m=+771.335223534" watchObservedRunningTime="2025-11-29 00:12:38.175704389 +0000 UTC m=+771.337597621" Nov 29 00:12:38 crc kubenswrapper[4931]: I1129 00:12:38.665027 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:38 crc kubenswrapper[4931]: I1129 00:12:38.665290 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:38 crc kubenswrapper[4931]: I1129 00:12:38.732439 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:39 crc kubenswrapper[4931]: I1129 00:12:39.248724 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:40 crc kubenswrapper[4931]: I1129 00:12:40.670742 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dkl9z"] Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.182713 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dkl9z" podUID="1228beae-7460-4619-afe9-2835885b5bb9" containerName="registry-server" containerID="cri-o://9c651e6acd56615f9b7f0f43fbd7b1df28ccf7897bb659658d4d71f6f93304bf" gracePeriod=2 Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.298764 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-cmcf4"] Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.299769 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-cmcf4" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.303658 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-8hprw" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.307599 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-cmcf4"] Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.311111 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6"] Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.311869 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.313986 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.326343 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-wv5fd"] Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.327306 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.333163 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6"] Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.418960 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4d93b5c8-a612-45f9-9351-31ffc5eb3b5a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jh9h6\" (UID: \"4d93b5c8-a612-45f9-9351-31ffc5eb3b5a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.419014 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8dt4\" (UniqueName: \"kubernetes.io/projected/e84ff97d-2d9a-4a4a-b619-a26e9c5a5595-kube-api-access-d8dt4\") pod \"nmstate-metrics-7f946cbc9-cmcf4\" (UID: \"e84ff97d-2d9a-4a4a-b619-a26e9c5a5595\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-cmcf4" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.419053 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7622d8c6-7baa-47e7-8f03-52ccd0d2f93f-ovs-socket\") pod \"nmstate-handler-wv5fd\" (UID: \"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f\") " pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.419197 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7622d8c6-7baa-47e7-8f03-52ccd0d2f93f-nmstate-lock\") pod \"nmstate-handler-wv5fd\" (UID: \"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f\") " pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.419324 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm4kk\" (UniqueName: \"kubernetes.io/projected/4d93b5c8-a612-45f9-9351-31ffc5eb3b5a-kube-api-access-wm4kk\") pod \"nmstate-webhook-5f6d4c5ccb-jh9h6\" (UID: \"4d93b5c8-a612-45f9-9351-31ffc5eb3b5a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.419389 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm"] Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.419403 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvdr7\" (UniqueName: \"kubernetes.io/projected/7622d8c6-7baa-47e7-8f03-52ccd0d2f93f-kube-api-access-hvdr7\") pod \"nmstate-handler-wv5fd\" (UID: \"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f\") " pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.419589 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7622d8c6-7baa-47e7-8f03-52ccd0d2f93f-dbus-socket\") pod \"nmstate-handler-wv5fd\" (UID: \"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f\") " pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.420141 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.421999 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.422018 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.422475 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-kjbr2" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.435406 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm"] Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.520480 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7622d8c6-7baa-47e7-8f03-52ccd0d2f93f-nmstate-lock\") pod \"nmstate-handler-wv5fd\" (UID: \"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f\") " pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.520609 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7622d8c6-7baa-47e7-8f03-52ccd0d2f93f-nmstate-lock\") pod \"nmstate-handler-wv5fd\" (UID: \"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f\") " pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.520843 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpdt7\" (UniqueName: \"kubernetes.io/projected/6fbf3473-ee9a-47b5-a782-47948ff5e590-kube-api-access-qpdt7\") pod \"nmstate-console-plugin-7fbb5f6569-g8dpm\" (UID: \"6fbf3473-ee9a-47b5-a782-47948ff5e590\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.520923 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm4kk\" (UniqueName: \"kubernetes.io/projected/4d93b5c8-a612-45f9-9351-31ffc5eb3b5a-kube-api-access-wm4kk\") pod \"nmstate-webhook-5f6d4c5ccb-jh9h6\" (UID: \"4d93b5c8-a612-45f9-9351-31ffc5eb3b5a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.520961 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvdr7\" (UniqueName: \"kubernetes.io/projected/7622d8c6-7baa-47e7-8f03-52ccd0d2f93f-kube-api-access-hvdr7\") pod \"nmstate-handler-wv5fd\" (UID: \"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f\") " pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.521012 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7622d8c6-7baa-47e7-8f03-52ccd0d2f93f-dbus-socket\") pod \"nmstate-handler-wv5fd\" (UID: \"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f\") " pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.521092 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8dt4\" (UniqueName: \"kubernetes.io/projected/e84ff97d-2d9a-4a4a-b619-a26e9c5a5595-kube-api-access-d8dt4\") pod \"nmstate-metrics-7f946cbc9-cmcf4\" (UID: \"e84ff97d-2d9a-4a4a-b619-a26e9c5a5595\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-cmcf4" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.521120 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4d93b5c8-a612-45f9-9351-31ffc5eb3b5a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jh9h6\" (UID: \"4d93b5c8-a612-45f9-9351-31ffc5eb3b5a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.521153 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7622d8c6-7baa-47e7-8f03-52ccd0d2f93f-ovs-socket\") pod \"nmstate-handler-wv5fd\" (UID: \"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f\") " pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.521208 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6fbf3473-ee9a-47b5-a782-47948ff5e590-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-g8dpm\" (UID: \"6fbf3473-ee9a-47b5-a782-47948ff5e590\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.521242 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fbf3473-ee9a-47b5-a782-47948ff5e590-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g8dpm\" (UID: \"6fbf3473-ee9a-47b5-a782-47948ff5e590\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.521326 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7622d8c6-7baa-47e7-8f03-52ccd0d2f93f-ovs-socket\") pod \"nmstate-handler-wv5fd\" (UID: \"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f\") " pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: E1129 00:12:42.521508 4931 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.521508 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7622d8c6-7baa-47e7-8f03-52ccd0d2f93f-dbus-socket\") pod \"nmstate-handler-wv5fd\" (UID: \"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f\") " pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: E1129 00:12:42.521641 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d93b5c8-a612-45f9-9351-31ffc5eb3b5a-tls-key-pair podName:4d93b5c8-a612-45f9-9351-31ffc5eb3b5a nodeName:}" failed. No retries permitted until 2025-11-29 00:12:43.02161617 +0000 UTC m=+776.183509462 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/4d93b5c8-a612-45f9-9351-31ffc5eb3b5a-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-jh9h6" (UID: "4d93b5c8-a612-45f9-9351-31ffc5eb3b5a") : secret "openshift-nmstate-webhook" not found Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.541173 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvdr7\" (UniqueName: \"kubernetes.io/projected/7622d8c6-7baa-47e7-8f03-52ccd0d2f93f-kube-api-access-hvdr7\") pod \"nmstate-handler-wv5fd\" (UID: \"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f\") " pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.541592 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8dt4\" (UniqueName: \"kubernetes.io/projected/e84ff97d-2d9a-4a4a-b619-a26e9c5a5595-kube-api-access-d8dt4\") pod \"nmstate-metrics-7f946cbc9-cmcf4\" (UID: \"e84ff97d-2d9a-4a4a-b619-a26e9c5a5595\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-cmcf4" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.553714 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm4kk\" (UniqueName: \"kubernetes.io/projected/4d93b5c8-a612-45f9-9351-31ffc5eb3b5a-kube-api-access-wm4kk\") pod \"nmstate-webhook-5f6d4c5ccb-jh9h6\" (UID: \"4d93b5c8-a612-45f9-9351-31ffc5eb3b5a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.617870 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-cmcf4" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.621833 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6fbf3473-ee9a-47b5-a782-47948ff5e590-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-g8dpm\" (UID: \"6fbf3473-ee9a-47b5-a782-47948ff5e590\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.621873 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fbf3473-ee9a-47b5-a782-47948ff5e590-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g8dpm\" (UID: \"6fbf3473-ee9a-47b5-a782-47948ff5e590\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.621909 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpdt7\" (UniqueName: \"kubernetes.io/projected/6fbf3473-ee9a-47b5-a782-47948ff5e590-kube-api-access-qpdt7\") pod \"nmstate-console-plugin-7fbb5f6569-g8dpm\" (UID: \"6fbf3473-ee9a-47b5-a782-47948ff5e590\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:42 crc kubenswrapper[4931]: E1129 00:12:42.622033 4931 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 29 00:12:42 crc kubenswrapper[4931]: E1129 00:12:42.622095 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fbf3473-ee9a-47b5-a782-47948ff5e590-plugin-serving-cert podName:6fbf3473-ee9a-47b5-a782-47948ff5e590 nodeName:}" failed. No retries permitted until 2025-11-29 00:12:43.122077532 +0000 UTC m=+776.283970764 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/6fbf3473-ee9a-47b5-a782-47948ff5e590-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-g8dpm" (UID: "6fbf3473-ee9a-47b5-a782-47948ff5e590") : secret "plugin-serving-cert" not found Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.622685 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6fbf3473-ee9a-47b5-a782-47948ff5e590-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-g8dpm\" (UID: \"6fbf3473-ee9a-47b5-a782-47948ff5e590\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.640390 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpdt7\" (UniqueName: \"kubernetes.io/projected/6fbf3473-ee9a-47b5-a782-47948ff5e590-kube-api-access-qpdt7\") pod \"nmstate-console-plugin-7fbb5f6569-g8dpm\" (UID: \"6fbf3473-ee9a-47b5-a782-47948ff5e590\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:42 crc kubenswrapper[4931]: I1129 00:12:42.645063 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.007104 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-cmcf4"] Nov 29 00:12:43 crc kubenswrapper[4931]: W1129 00:12:43.016170 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode84ff97d_2d9a_4a4a_b619_a26e9c5a5595.slice/crio-22b5b24b91b8fc82b9ac4fc9e027398946ce4da0d7e6bae0dd1bd24f8d946d0e WatchSource:0}: Error finding container 22b5b24b91b8fc82b9ac4fc9e027398946ce4da0d7e6bae0dd1bd24f8d946d0e: Status 404 returned error can't find the container with id 22b5b24b91b8fc82b9ac4fc9e027398946ce4da0d7e6bae0dd1bd24f8d946d0e Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.031419 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4d93b5c8-a612-45f9-9351-31ffc5eb3b5a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jh9h6\" (UID: \"4d93b5c8-a612-45f9-9351-31ffc5eb3b5a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" Nov 29 00:12:43 crc kubenswrapper[4931]: E1129 00:12:43.031582 4931 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 29 00:12:43 crc kubenswrapper[4931]: E1129 00:12:43.031639 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d93b5c8-a612-45f9-9351-31ffc5eb3b5a-tls-key-pair podName:4d93b5c8-a612-45f9-9351-31ffc5eb3b5a nodeName:}" failed. No retries permitted until 2025-11-29 00:12:44.031624229 +0000 UTC m=+777.193517461 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/4d93b5c8-a612-45f9-9351-31ffc5eb3b5a-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-jh9h6" (UID: "4d93b5c8-a612-45f9-9351-31ffc5eb3b5a") : secret "openshift-nmstate-webhook" not found Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.132392 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fbf3473-ee9a-47b5-a782-47948ff5e590-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g8dpm\" (UID: \"6fbf3473-ee9a-47b5-a782-47948ff5e590\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:43 crc kubenswrapper[4931]: E1129 00:12:43.132641 4931 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 29 00:12:43 crc kubenswrapper[4931]: E1129 00:12:43.132737 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fbf3473-ee9a-47b5-a782-47948ff5e590-plugin-serving-cert podName:6fbf3473-ee9a-47b5-a782-47948ff5e590 nodeName:}" failed. No retries permitted until 2025-11-29 00:12:44.1327154 +0000 UTC m=+777.294608652 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/6fbf3473-ee9a-47b5-a782-47948ff5e590-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-g8dpm" (UID: "6fbf3473-ee9a-47b5-a782-47948ff5e590") : secret "plugin-serving-cert" not found Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.192611 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-wv5fd" event={"ID":"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f","Type":"ContainerStarted","Data":"3dbfdc86c42f812120ef8a834e83bd124929d9f2216b88e3e2e8dbcd2579168b"} Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.193899 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-cmcf4" event={"ID":"e84ff97d-2d9a-4a4a-b619-a26e9c5a5595","Type":"ContainerStarted","Data":"22b5b24b91b8fc82b9ac4fc9e027398946ce4da0d7e6bae0dd1bd24f8d946d0e"} Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.698397 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-78df5467d-7cj2q"] Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.699249 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.718140 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-78df5467d-7cj2q"] Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.762327 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/be4a5691-971c-4478-aa77-f223f7aafa6c-oauth-serving-cert\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.762379 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/be4a5691-971c-4478-aa77-f223f7aafa6c-console-oauth-config\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.762421 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/be4a5691-971c-4478-aa77-f223f7aafa6c-console-config\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.762442 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/be4a5691-971c-4478-aa77-f223f7aafa6c-service-ca\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.762515 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk9km\" (UniqueName: \"kubernetes.io/projected/be4a5691-971c-4478-aa77-f223f7aafa6c-kube-api-access-vk9km\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.762550 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/be4a5691-971c-4478-aa77-f223f7aafa6c-console-serving-cert\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.762618 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be4a5691-971c-4478-aa77-f223f7aafa6c-trusted-ca-bundle\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.863843 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk9km\" (UniqueName: \"kubernetes.io/projected/be4a5691-971c-4478-aa77-f223f7aafa6c-kube-api-access-vk9km\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.863887 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/be4a5691-971c-4478-aa77-f223f7aafa6c-console-serving-cert\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.863925 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be4a5691-971c-4478-aa77-f223f7aafa6c-trusted-ca-bundle\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.863951 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/be4a5691-971c-4478-aa77-f223f7aafa6c-oauth-serving-cert\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.863970 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/be4a5691-971c-4478-aa77-f223f7aafa6c-console-oauth-config\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.863989 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/be4a5691-971c-4478-aa77-f223f7aafa6c-console-config\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.864005 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/be4a5691-971c-4478-aa77-f223f7aafa6c-service-ca\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.864947 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/be4a5691-971c-4478-aa77-f223f7aafa6c-service-ca\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.865145 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/be4a5691-971c-4478-aa77-f223f7aafa6c-oauth-serving-cert\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.865260 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/be4a5691-971c-4478-aa77-f223f7aafa6c-console-config\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.865430 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be4a5691-971c-4478-aa77-f223f7aafa6c-trusted-ca-bundle\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.869513 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/be4a5691-971c-4478-aa77-f223f7aafa6c-console-serving-cert\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.871367 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/be4a5691-971c-4478-aa77-f223f7aafa6c-console-oauth-config\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:43 crc kubenswrapper[4931]: I1129 00:12:43.882906 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk9km\" (UniqueName: \"kubernetes.io/projected/be4a5691-971c-4478-aa77-f223f7aafa6c-kube-api-access-vk9km\") pod \"console-78df5467d-7cj2q\" (UID: \"be4a5691-971c-4478-aa77-f223f7aafa6c\") " pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.065626 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4d93b5c8-a612-45f9-9351-31ffc5eb3b5a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jh9h6\" (UID: \"4d93b5c8-a612-45f9-9351-31ffc5eb3b5a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.068282 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.078463 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4d93b5c8-a612-45f9-9351-31ffc5eb3b5a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jh9h6\" (UID: \"4d93b5c8-a612-45f9-9351-31ffc5eb3b5a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.137950 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.167175 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fbf3473-ee9a-47b5-a782-47948ff5e590-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g8dpm\" (UID: \"6fbf3473-ee9a-47b5-a782-47948ff5e590\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.173643 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fbf3473-ee9a-47b5-a782-47948ff5e590-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g8dpm\" (UID: \"6fbf3473-ee9a-47b5-a782-47948ff5e590\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.204985 4931 generic.go:334] "Generic (PLEG): container finished" podID="1228beae-7460-4619-afe9-2835885b5bb9" containerID="9c651e6acd56615f9b7f0f43fbd7b1df28ccf7897bb659658d4d71f6f93304bf" exitCode=0 Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.205053 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dkl9z" event={"ID":"1228beae-7460-4619-afe9-2835885b5bb9","Type":"ContainerDied","Data":"9c651e6acd56615f9b7f0f43fbd7b1df28ccf7897bb659658d4d71f6f93304bf"} Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.233205 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.296713 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-78df5467d-7cj2q"] Nov 29 00:12:44 crc kubenswrapper[4931]: W1129 00:12:44.306569 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe4a5691_971c_4478_aa77_f223f7aafa6c.slice/crio-8716c3ac9ce707ceb33133a110aa3ebde5b4728ad2f33875af08aa9692f8c454 WatchSource:0}: Error finding container 8716c3ac9ce707ceb33133a110aa3ebde5b4728ad2f33875af08aa9692f8c454: Status 404 returned error can't find the container with id 8716c3ac9ce707ceb33133a110aa3ebde5b4728ad2f33875af08aa9692f8c454 Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.506831 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm"] Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.515892 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.564711 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6"] Nov 29 00:12:44 crc kubenswrapper[4931]: W1129 00:12:44.579787 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d93b5c8_a612_45f9_9351_31ffc5eb3b5a.slice/crio-e4392a808e62d47958d498a2a3bc4ac6af35b3d8f729da7eef732d69f8077aa4 WatchSource:0}: Error finding container e4392a808e62d47958d498a2a3bc4ac6af35b3d8f729da7eef732d69f8077aa4: Status 404 returned error can't find the container with id e4392a808e62d47958d498a2a3bc4ac6af35b3d8f729da7eef732d69f8077aa4 Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.674966 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1228beae-7460-4619-afe9-2835885b5bb9-catalog-content\") pod \"1228beae-7460-4619-afe9-2835885b5bb9\" (UID: \"1228beae-7460-4619-afe9-2835885b5bb9\") " Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.675026 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-467n7\" (UniqueName: \"kubernetes.io/projected/1228beae-7460-4619-afe9-2835885b5bb9-kube-api-access-467n7\") pod \"1228beae-7460-4619-afe9-2835885b5bb9\" (UID: \"1228beae-7460-4619-afe9-2835885b5bb9\") " Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.675143 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1228beae-7460-4619-afe9-2835885b5bb9-utilities\") pod \"1228beae-7460-4619-afe9-2835885b5bb9\" (UID: \"1228beae-7460-4619-afe9-2835885b5bb9\") " Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.676236 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1228beae-7460-4619-afe9-2835885b5bb9-utilities" (OuterVolumeSpecName: "utilities") pod "1228beae-7460-4619-afe9-2835885b5bb9" (UID: "1228beae-7460-4619-afe9-2835885b5bb9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.684343 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1228beae-7460-4619-afe9-2835885b5bb9-kube-api-access-467n7" (OuterVolumeSpecName: "kube-api-access-467n7") pod "1228beae-7460-4619-afe9-2835885b5bb9" (UID: "1228beae-7460-4619-afe9-2835885b5bb9"). InnerVolumeSpecName "kube-api-access-467n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.771301 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1228beae-7460-4619-afe9-2835885b5bb9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1228beae-7460-4619-afe9-2835885b5bb9" (UID: "1228beae-7460-4619-afe9-2835885b5bb9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.776223 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1228beae-7460-4619-afe9-2835885b5bb9-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.776252 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1228beae-7460-4619-afe9-2835885b5bb9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:12:44 crc kubenswrapper[4931]: I1129 00:12:44.776263 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-467n7\" (UniqueName: \"kubernetes.io/projected/1228beae-7460-4619-afe9-2835885b5bb9-kube-api-access-467n7\") on node \"crc\" DevicePath \"\"" Nov 29 00:12:45 crc kubenswrapper[4931]: I1129 00:12:45.218541 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dkl9z" Nov 29 00:12:45 crc kubenswrapper[4931]: I1129 00:12:45.225135 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-78df5467d-7cj2q" event={"ID":"be4a5691-971c-4478-aa77-f223f7aafa6c","Type":"ContainerStarted","Data":"14a82158bc24756e8604ff7f806ebfc984f84e9f62309b8bcdd493c8d86725b2"} Nov 29 00:12:45 crc kubenswrapper[4931]: I1129 00:12:45.225180 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-78df5467d-7cj2q" event={"ID":"be4a5691-971c-4478-aa77-f223f7aafa6c","Type":"ContainerStarted","Data":"8716c3ac9ce707ceb33133a110aa3ebde5b4728ad2f33875af08aa9692f8c454"} Nov 29 00:12:45 crc kubenswrapper[4931]: I1129 00:12:45.225196 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dkl9z" event={"ID":"1228beae-7460-4619-afe9-2835885b5bb9","Type":"ContainerDied","Data":"f9a4a152e3bb981addd803e67e81f793c1cbdcc66101acd218485cf97e38cba3"} Nov 29 00:12:45 crc kubenswrapper[4931]: I1129 00:12:45.225320 4931 scope.go:117] "RemoveContainer" containerID="9c651e6acd56615f9b7f0f43fbd7b1df28ccf7897bb659658d4d71f6f93304bf" Nov 29 00:12:45 crc kubenswrapper[4931]: I1129 00:12:45.226209 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" event={"ID":"4d93b5c8-a612-45f9-9351-31ffc5eb3b5a","Type":"ContainerStarted","Data":"e4392a808e62d47958d498a2a3bc4ac6af35b3d8f729da7eef732d69f8077aa4"} Nov 29 00:12:45 crc kubenswrapper[4931]: I1129 00:12:45.229930 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-78df5467d-7cj2q" podStartSLOduration=2.229912704 podStartE2EDuration="2.229912704s" podCreationTimestamp="2025-11-29 00:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:12:45.227904077 +0000 UTC m=+778.389797329" watchObservedRunningTime="2025-11-29 00:12:45.229912704 +0000 UTC m=+778.391805926" Nov 29 00:12:45 crc kubenswrapper[4931]: I1129 00:12:45.230261 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" event={"ID":"6fbf3473-ee9a-47b5-a782-47948ff5e590","Type":"ContainerStarted","Data":"e0452ab1ca0330872e0082bedd077cd7aa6967a0b3dc15ddb8d46b7f7591bbb4"} Nov 29 00:12:45 crc kubenswrapper[4931]: I1129 00:12:45.250231 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dkl9z"] Nov 29 00:12:45 crc kubenswrapper[4931]: I1129 00:12:45.253599 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dkl9z"] Nov 29 00:12:46 crc kubenswrapper[4931]: I1129 00:12:46.008871 4931 scope.go:117] "RemoveContainer" containerID="d62d0c6a4bdb0ab082f33a5d1fdd2364d63088b33f81e777788673066b0c7718" Nov 29 00:12:46 crc kubenswrapper[4931]: I1129 00:12:46.036884 4931 scope.go:117] "RemoveContainer" containerID="f9be5104476974554468c119b3ea875f5231c954e350ec243a0f7b4c50e25a30" Nov 29 00:12:47 crc kubenswrapper[4931]: I1129 00:12:47.226203 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1228beae-7460-4619-afe9-2835885b5bb9" path="/var/lib/kubelet/pods/1228beae-7460-4619-afe9-2835885b5bb9/volumes" Nov 29 00:12:47 crc kubenswrapper[4931]: I1129 00:12:47.249087 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-wv5fd" event={"ID":"7622d8c6-7baa-47e7-8f03-52ccd0d2f93f","Type":"ContainerStarted","Data":"d4832ccd45a67f330826d10ff45ca657cfeb12107aa9c1db6fef8c2a130a4770"} Nov 29 00:12:47 crc kubenswrapper[4931]: I1129 00:12:47.249407 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:47 crc kubenswrapper[4931]: I1129 00:12:47.251614 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" event={"ID":"4d93b5c8-a612-45f9-9351-31ffc5eb3b5a","Type":"ContainerStarted","Data":"d26b940863b78d62f9e9d4e3d0d889af3edf8f7df7503a73bf37f30e37017771"} Nov 29 00:12:47 crc kubenswrapper[4931]: I1129 00:12:47.251775 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" Nov 29 00:12:47 crc kubenswrapper[4931]: I1129 00:12:47.254354 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-cmcf4" event={"ID":"e84ff97d-2d9a-4a4a-b619-a26e9c5a5595","Type":"ContainerStarted","Data":"bd4ae05f99f7d3fac97151a0cba770fa26a0bd6e9d503e5476bca92b99b5ac1c"} Nov 29 00:12:47 crc kubenswrapper[4931]: I1129 00:12:47.306881 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" podStartSLOduration=3.739745112 podStartE2EDuration="5.306840848s" podCreationTimestamp="2025-11-29 00:12:42 +0000 UTC" firstStartedPulling="2025-11-29 00:12:44.582618342 +0000 UTC m=+777.744511574" lastFinishedPulling="2025-11-29 00:12:46.149714068 +0000 UTC m=+779.311607310" observedRunningTime="2025-11-29 00:12:47.306572552 +0000 UTC m=+780.468465854" watchObservedRunningTime="2025-11-29 00:12:47.306840848 +0000 UTC m=+780.468734090" Nov 29 00:12:47 crc kubenswrapper[4931]: I1129 00:12:47.341667 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-wv5fd" podStartSLOduration=1.873581797 podStartE2EDuration="5.34164782s" podCreationTimestamp="2025-11-29 00:12:42 +0000 UTC" firstStartedPulling="2025-11-29 00:12:42.677866215 +0000 UTC m=+775.839759447" lastFinishedPulling="2025-11-29 00:12:46.145932218 +0000 UTC m=+779.307825470" observedRunningTime="2025-11-29 00:12:47.333630558 +0000 UTC m=+780.495523830" watchObservedRunningTime="2025-11-29 00:12:47.34164782 +0000 UTC m=+780.503541062" Nov 29 00:12:50 crc kubenswrapper[4931]: I1129 00:12:50.286178 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-cmcf4" event={"ID":"e84ff97d-2d9a-4a4a-b619-a26e9c5a5595","Type":"ContainerStarted","Data":"1fba8606f0b5653fc594a3fbae09f6fe082b3b3b95b8afdd9a406ba29a8bf9f5"} Nov 29 00:12:50 crc kubenswrapper[4931]: I1129 00:12:50.287792 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" event={"ID":"6fbf3473-ee9a-47b5-a782-47948ff5e590","Type":"ContainerStarted","Data":"e2b1222cbf6f69e9e6550a3ef0ef4525a51a6c68b1b3f45ce85af7e489df5629"} Nov 29 00:12:50 crc kubenswrapper[4931]: I1129 00:12:50.310804 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-cmcf4" podStartSLOduration=1.806718729 podStartE2EDuration="8.310771864s" podCreationTimestamp="2025-11-29 00:12:42 +0000 UTC" firstStartedPulling="2025-11-29 00:12:43.019027154 +0000 UTC m=+776.180920396" lastFinishedPulling="2025-11-29 00:12:49.523080299 +0000 UTC m=+782.684973531" observedRunningTime="2025-11-29 00:12:50.304640257 +0000 UTC m=+783.466533509" watchObservedRunningTime="2025-11-29 00:12:50.310771864 +0000 UTC m=+783.472665126" Nov 29 00:12:50 crc kubenswrapper[4931]: I1129 00:12:50.328719 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g8dpm" podStartSLOduration=3.404500724 podStartE2EDuration="8.328694482s" podCreationTimestamp="2025-11-29 00:12:42 +0000 UTC" firstStartedPulling="2025-11-29 00:12:44.547741049 +0000 UTC m=+777.709634281" lastFinishedPulling="2025-11-29 00:12:49.471934807 +0000 UTC m=+782.633828039" observedRunningTime="2025-11-29 00:12:50.32440959 +0000 UTC m=+783.486302852" watchObservedRunningTime="2025-11-29 00:12:50.328694482 +0000 UTC m=+783.490587744" Nov 29 00:12:52 crc kubenswrapper[4931]: I1129 00:12:52.686166 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-wv5fd" Nov 29 00:12:53 crc kubenswrapper[4931]: I1129 00:12:53.735501 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:12:53 crc kubenswrapper[4931]: I1129 00:12:53.735931 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:12:54 crc kubenswrapper[4931]: I1129 00:12:54.068915 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:54 crc kubenswrapper[4931]: I1129 00:12:54.069147 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:54 crc kubenswrapper[4931]: I1129 00:12:54.079168 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:54 crc kubenswrapper[4931]: I1129 00:12:54.323416 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-78df5467d-7cj2q" Nov 29 00:12:54 crc kubenswrapper[4931]: I1129 00:12:54.395159 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vvmnt"] Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.257396 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kd4xm"] Nov 29 00:13:02 crc kubenswrapper[4931]: E1129 00:13:02.259849 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1228beae-7460-4619-afe9-2835885b5bb9" containerName="extract-content" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.260017 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1228beae-7460-4619-afe9-2835885b5bb9" containerName="extract-content" Nov 29 00:13:02 crc kubenswrapper[4931]: E1129 00:13:02.260130 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1228beae-7460-4619-afe9-2835885b5bb9" containerName="registry-server" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.260214 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1228beae-7460-4619-afe9-2835885b5bb9" containerName="registry-server" Nov 29 00:13:02 crc kubenswrapper[4931]: E1129 00:13:02.260303 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1228beae-7460-4619-afe9-2835885b5bb9" containerName="extract-utilities" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.260378 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1228beae-7460-4619-afe9-2835885b5bb9" containerName="extract-utilities" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.260637 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="1228beae-7460-4619-afe9-2835885b5bb9" containerName="registry-server" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.261990 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.276409 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd4xm"] Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.321266 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92fkr\" (UniqueName: \"kubernetes.io/projected/524a1733-89a9-4cff-82cf-55b342f48037-kube-api-access-92fkr\") pod \"redhat-marketplace-kd4xm\" (UID: \"524a1733-89a9-4cff-82cf-55b342f48037\") " pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.321488 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524a1733-89a9-4cff-82cf-55b342f48037-catalog-content\") pod \"redhat-marketplace-kd4xm\" (UID: \"524a1733-89a9-4cff-82cf-55b342f48037\") " pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.321705 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524a1733-89a9-4cff-82cf-55b342f48037-utilities\") pod \"redhat-marketplace-kd4xm\" (UID: \"524a1733-89a9-4cff-82cf-55b342f48037\") " pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.422606 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92fkr\" (UniqueName: \"kubernetes.io/projected/524a1733-89a9-4cff-82cf-55b342f48037-kube-api-access-92fkr\") pod \"redhat-marketplace-kd4xm\" (UID: \"524a1733-89a9-4cff-82cf-55b342f48037\") " pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.422667 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524a1733-89a9-4cff-82cf-55b342f48037-catalog-content\") pod \"redhat-marketplace-kd4xm\" (UID: \"524a1733-89a9-4cff-82cf-55b342f48037\") " pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.422692 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524a1733-89a9-4cff-82cf-55b342f48037-utilities\") pod \"redhat-marketplace-kd4xm\" (UID: \"524a1733-89a9-4cff-82cf-55b342f48037\") " pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.423179 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524a1733-89a9-4cff-82cf-55b342f48037-utilities\") pod \"redhat-marketplace-kd4xm\" (UID: \"524a1733-89a9-4cff-82cf-55b342f48037\") " pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.423450 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524a1733-89a9-4cff-82cf-55b342f48037-catalog-content\") pod \"redhat-marketplace-kd4xm\" (UID: \"524a1733-89a9-4cff-82cf-55b342f48037\") " pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.447236 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92fkr\" (UniqueName: \"kubernetes.io/projected/524a1733-89a9-4cff-82cf-55b342f48037-kube-api-access-92fkr\") pod \"redhat-marketplace-kd4xm\" (UID: \"524a1733-89a9-4cff-82cf-55b342f48037\") " pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:02 crc kubenswrapper[4931]: I1129 00:13:02.599067 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:03 crc kubenswrapper[4931]: I1129 00:13:03.048500 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd4xm"] Nov 29 00:13:03 crc kubenswrapper[4931]: I1129 00:13:03.379727 4931 generic.go:334] "Generic (PLEG): container finished" podID="524a1733-89a9-4cff-82cf-55b342f48037" containerID="1213923a67320c7b2647370a4321ad0a444fd0964f0e634d1349abed43fa858d" exitCode=0 Nov 29 00:13:03 crc kubenswrapper[4931]: I1129 00:13:03.379962 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd4xm" event={"ID":"524a1733-89a9-4cff-82cf-55b342f48037","Type":"ContainerDied","Data":"1213923a67320c7b2647370a4321ad0a444fd0964f0e634d1349abed43fa858d"} Nov 29 00:13:03 crc kubenswrapper[4931]: I1129 00:13:03.380928 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd4xm" event={"ID":"524a1733-89a9-4cff-82cf-55b342f48037","Type":"ContainerStarted","Data":"04b2e9c043ebb38ef401e1520ec71400bbebf2a4a62268baa147aaac30f44474"} Nov 29 00:13:04 crc kubenswrapper[4931]: I1129 00:13:04.146999 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jh9h6" Nov 29 00:13:04 crc kubenswrapper[4931]: I1129 00:13:04.390330 4931 generic.go:334] "Generic (PLEG): container finished" podID="524a1733-89a9-4cff-82cf-55b342f48037" containerID="c01fb8b1accb0ed282f49516a48e0529d3abfbe68947508eb033f9bd4ebf55a4" exitCode=0 Nov 29 00:13:04 crc kubenswrapper[4931]: I1129 00:13:04.390434 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd4xm" event={"ID":"524a1733-89a9-4cff-82cf-55b342f48037","Type":"ContainerDied","Data":"c01fb8b1accb0ed282f49516a48e0529d3abfbe68947508eb033f9bd4ebf55a4"} Nov 29 00:13:05 crc kubenswrapper[4931]: I1129 00:13:05.401105 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd4xm" event={"ID":"524a1733-89a9-4cff-82cf-55b342f48037","Type":"ContainerStarted","Data":"b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675"} Nov 29 00:13:05 crc kubenswrapper[4931]: I1129 00:13:05.420884 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kd4xm" podStartSLOduration=1.97013536 podStartE2EDuration="3.420860413s" podCreationTimestamp="2025-11-29 00:13:02 +0000 UTC" firstStartedPulling="2025-11-29 00:13:03.38237173 +0000 UTC m=+796.544264992" lastFinishedPulling="2025-11-29 00:13:04.833096803 +0000 UTC m=+797.994990045" observedRunningTime="2025-11-29 00:13:05.419906041 +0000 UTC m=+798.581799283" watchObservedRunningTime="2025-11-29 00:13:05.420860413 +0000 UTC m=+798.582753655" Nov 29 00:13:12 crc kubenswrapper[4931]: I1129 00:13:12.600195 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:12 crc kubenswrapper[4931]: I1129 00:13:12.600778 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:12 crc kubenswrapper[4931]: I1129 00:13:12.658164 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:13 crc kubenswrapper[4931]: I1129 00:13:13.510757 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:13 crc kubenswrapper[4931]: I1129 00:13:13.556186 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd4xm"] Nov 29 00:13:15 crc kubenswrapper[4931]: I1129 00:13:15.462754 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kd4xm" podUID="524a1733-89a9-4cff-82cf-55b342f48037" containerName="registry-server" containerID="cri-o://b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675" gracePeriod=2 Nov 29 00:13:16 crc kubenswrapper[4931]: I1129 00:13:16.996620 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.028331 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524a1733-89a9-4cff-82cf-55b342f48037-catalog-content\") pod \"524a1733-89a9-4cff-82cf-55b342f48037\" (UID: \"524a1733-89a9-4cff-82cf-55b342f48037\") " Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.028736 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92fkr\" (UniqueName: \"kubernetes.io/projected/524a1733-89a9-4cff-82cf-55b342f48037-kube-api-access-92fkr\") pod \"524a1733-89a9-4cff-82cf-55b342f48037\" (UID: \"524a1733-89a9-4cff-82cf-55b342f48037\") " Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.028778 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524a1733-89a9-4cff-82cf-55b342f48037-utilities\") pod \"524a1733-89a9-4cff-82cf-55b342f48037\" (UID: \"524a1733-89a9-4cff-82cf-55b342f48037\") " Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.030417 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/524a1733-89a9-4cff-82cf-55b342f48037-utilities" (OuterVolumeSpecName: "utilities") pod "524a1733-89a9-4cff-82cf-55b342f48037" (UID: "524a1733-89a9-4cff-82cf-55b342f48037"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.034459 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/524a1733-89a9-4cff-82cf-55b342f48037-kube-api-access-92fkr" (OuterVolumeSpecName: "kube-api-access-92fkr") pod "524a1733-89a9-4cff-82cf-55b342f48037" (UID: "524a1733-89a9-4cff-82cf-55b342f48037"). InnerVolumeSpecName "kube-api-access-92fkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.051985 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/524a1733-89a9-4cff-82cf-55b342f48037-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "524a1733-89a9-4cff-82cf-55b342f48037" (UID: "524a1733-89a9-4cff-82cf-55b342f48037"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.131111 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92fkr\" (UniqueName: \"kubernetes.io/projected/524a1733-89a9-4cff-82cf-55b342f48037-kube-api-access-92fkr\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.131338 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/524a1733-89a9-4cff-82cf-55b342f48037-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.131398 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/524a1733-89a9-4cff-82cf-55b342f48037-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.479853 4931 generic.go:334] "Generic (PLEG): container finished" podID="524a1733-89a9-4cff-82cf-55b342f48037" containerID="b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675" exitCode=0 Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.479915 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd4xm" event={"ID":"524a1733-89a9-4cff-82cf-55b342f48037","Type":"ContainerDied","Data":"b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675"} Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.479973 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd4xm" event={"ID":"524a1733-89a9-4cff-82cf-55b342f48037","Type":"ContainerDied","Data":"04b2e9c043ebb38ef401e1520ec71400bbebf2a4a62268baa147aaac30f44474"} Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.479993 4931 scope.go:117] "RemoveContainer" containerID="b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.480003 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd4xm" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.505741 4931 scope.go:117] "RemoveContainer" containerID="c01fb8b1accb0ed282f49516a48e0529d3abfbe68947508eb033f9bd4ebf55a4" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.516792 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd4xm"] Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.521841 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd4xm"] Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.527968 4931 scope.go:117] "RemoveContainer" containerID="1213923a67320c7b2647370a4321ad0a444fd0964f0e634d1349abed43fa858d" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.567531 4931 scope.go:117] "RemoveContainer" containerID="b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675" Nov 29 00:13:17 crc kubenswrapper[4931]: E1129 00:13:17.568038 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675\": container with ID starting with b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675 not found: ID does not exist" containerID="b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.568095 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675"} err="failed to get container status \"b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675\": rpc error: code = NotFound desc = could not find container \"b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675\": container with ID starting with b8f3462c5292a11fa5d5af48ffec0cbb9ad7325ef479dbf1b1cb299334c57675 not found: ID does not exist" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.568126 4931 scope.go:117] "RemoveContainer" containerID="c01fb8b1accb0ed282f49516a48e0529d3abfbe68947508eb033f9bd4ebf55a4" Nov 29 00:13:17 crc kubenswrapper[4931]: E1129 00:13:17.568411 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c01fb8b1accb0ed282f49516a48e0529d3abfbe68947508eb033f9bd4ebf55a4\": container with ID starting with c01fb8b1accb0ed282f49516a48e0529d3abfbe68947508eb033f9bd4ebf55a4 not found: ID does not exist" containerID="c01fb8b1accb0ed282f49516a48e0529d3abfbe68947508eb033f9bd4ebf55a4" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.568446 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c01fb8b1accb0ed282f49516a48e0529d3abfbe68947508eb033f9bd4ebf55a4"} err="failed to get container status \"c01fb8b1accb0ed282f49516a48e0529d3abfbe68947508eb033f9bd4ebf55a4\": rpc error: code = NotFound desc = could not find container \"c01fb8b1accb0ed282f49516a48e0529d3abfbe68947508eb033f9bd4ebf55a4\": container with ID starting with c01fb8b1accb0ed282f49516a48e0529d3abfbe68947508eb033f9bd4ebf55a4 not found: ID does not exist" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.568491 4931 scope.go:117] "RemoveContainer" containerID="1213923a67320c7b2647370a4321ad0a444fd0964f0e634d1349abed43fa858d" Nov 29 00:13:17 crc kubenswrapper[4931]: E1129 00:13:17.568755 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1213923a67320c7b2647370a4321ad0a444fd0964f0e634d1349abed43fa858d\": container with ID starting with 1213923a67320c7b2647370a4321ad0a444fd0964f0e634d1349abed43fa858d not found: ID does not exist" containerID="1213923a67320c7b2647370a4321ad0a444fd0964f0e634d1349abed43fa858d" Nov 29 00:13:17 crc kubenswrapper[4931]: I1129 00:13:17.568776 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1213923a67320c7b2647370a4321ad0a444fd0964f0e634d1349abed43fa858d"} err="failed to get container status \"1213923a67320c7b2647370a4321ad0a444fd0964f0e634d1349abed43fa858d\": rpc error: code = NotFound desc = could not find container \"1213923a67320c7b2647370a4321ad0a444fd0964f0e634d1349abed43fa858d\": container with ID starting with 1213923a67320c7b2647370a4321ad0a444fd0964f0e634d1349abed43fa858d not found: ID does not exist" Nov 29 00:13:18 crc kubenswrapper[4931]: I1129 00:13:18.887694 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44"] Nov 29 00:13:18 crc kubenswrapper[4931]: E1129 00:13:18.888658 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524a1733-89a9-4cff-82cf-55b342f48037" containerName="extract-content" Nov 29 00:13:18 crc kubenswrapper[4931]: I1129 00:13:18.888689 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="524a1733-89a9-4cff-82cf-55b342f48037" containerName="extract-content" Nov 29 00:13:18 crc kubenswrapper[4931]: E1129 00:13:18.888713 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524a1733-89a9-4cff-82cf-55b342f48037" containerName="registry-server" Nov 29 00:13:18 crc kubenswrapper[4931]: I1129 00:13:18.888729 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="524a1733-89a9-4cff-82cf-55b342f48037" containerName="registry-server" Nov 29 00:13:18 crc kubenswrapper[4931]: E1129 00:13:18.888759 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524a1733-89a9-4cff-82cf-55b342f48037" containerName="extract-utilities" Nov 29 00:13:18 crc kubenswrapper[4931]: I1129 00:13:18.888778 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="524a1733-89a9-4cff-82cf-55b342f48037" containerName="extract-utilities" Nov 29 00:13:18 crc kubenswrapper[4931]: I1129 00:13:18.889017 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="524a1733-89a9-4cff-82cf-55b342f48037" containerName="registry-server" Nov 29 00:13:18 crc kubenswrapper[4931]: I1129 00:13:18.890726 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:18 crc kubenswrapper[4931]: I1129 00:13:18.895980 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 29 00:13:18 crc kubenswrapper[4931]: I1129 00:13:18.896875 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44"] Nov 29 00:13:18 crc kubenswrapper[4931]: I1129 00:13:18.964558 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86b012aa-97b4-461d-abf9-19a1a261b806-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44\" (UID: \"86b012aa-97b4-461d-abf9-19a1a261b806\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:18 crc kubenswrapper[4931]: I1129 00:13:18.964677 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86b012aa-97b4-461d-abf9-19a1a261b806-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44\" (UID: \"86b012aa-97b4-461d-abf9-19a1a261b806\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:18 crc kubenswrapper[4931]: I1129 00:13:18.964732 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6ffk\" (UniqueName: \"kubernetes.io/projected/86b012aa-97b4-461d-abf9-19a1a261b806-kube-api-access-b6ffk\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44\" (UID: \"86b012aa-97b4-461d-abf9-19a1a261b806\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.065882 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6ffk\" (UniqueName: \"kubernetes.io/projected/86b012aa-97b4-461d-abf9-19a1a261b806-kube-api-access-b6ffk\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44\" (UID: \"86b012aa-97b4-461d-abf9-19a1a261b806\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.066006 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86b012aa-97b4-461d-abf9-19a1a261b806-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44\" (UID: \"86b012aa-97b4-461d-abf9-19a1a261b806\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.066042 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86b012aa-97b4-461d-abf9-19a1a261b806-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44\" (UID: \"86b012aa-97b4-461d-abf9-19a1a261b806\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.066589 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86b012aa-97b4-461d-abf9-19a1a261b806-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44\" (UID: \"86b012aa-97b4-461d-abf9-19a1a261b806\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.066750 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86b012aa-97b4-461d-abf9-19a1a261b806-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44\" (UID: \"86b012aa-97b4-461d-abf9-19a1a261b806\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.085760 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6ffk\" (UniqueName: \"kubernetes.io/projected/86b012aa-97b4-461d-abf9-19a1a261b806-kube-api-access-b6ffk\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44\" (UID: \"86b012aa-97b4-461d-abf9-19a1a261b806\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.224295 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="524a1733-89a9-4cff-82cf-55b342f48037" path="/var/lib/kubelet/pods/524a1733-89a9-4cff-82cf-55b342f48037/volumes" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.249760 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.438454 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-vvmnt" podUID="2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" containerName="console" containerID="cri-o://160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd" gracePeriod=15 Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.507930 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44"] Nov 29 00:13:19 crc kubenswrapper[4931]: W1129 00:13:19.531480 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86b012aa_97b4_461d_abf9_19a1a261b806.slice/crio-c6ba17ba3fb432e150d6e1b760f2f0a28795b6b4280db231a86cce95603dba17 WatchSource:0}: Error finding container c6ba17ba3fb432e150d6e1b760f2f0a28795b6b4280db231a86cce95603dba17: Status 404 returned error can't find the container with id c6ba17ba3fb432e150d6e1b760f2f0a28795b6b4280db231a86cce95603dba17 Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.838956 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vvmnt_2f63c980-6d60-4de2-8b1a-6a3e5d1b3864/console/0.log" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.839028 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.981280 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-serving-cert\") pod \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.981379 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-config\") pod \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.981420 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-service-ca\") pod \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.981477 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-oauth-serving-cert\") pod \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.981539 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-trusted-ca-bundle\") pod \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.981572 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86nlr\" (UniqueName: \"kubernetes.io/projected/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-kube-api-access-86nlr\") pod \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.981607 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-oauth-config\") pod \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\" (UID: \"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864\") " Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.982539 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" (UID: "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.982767 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-service-ca" (OuterVolumeSpecName: "service-ca") pod "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" (UID: "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.982959 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" (UID: "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.983064 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-config" (OuterVolumeSpecName: "console-config") pod "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" (UID: "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.987514 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" (UID: "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.987878 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-kube-api-access-86nlr" (OuterVolumeSpecName: "kube-api-access-86nlr") pod "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" (UID: "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864"). InnerVolumeSpecName "kube-api-access-86nlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:13:19 crc kubenswrapper[4931]: I1129 00:13:19.988513 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" (UID: "2f63c980-6d60-4de2-8b1a-6a3e5d1b3864"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.082800 4931 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.082877 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86nlr\" (UniqueName: \"kubernetes.io/projected/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-kube-api-access-86nlr\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.082902 4931 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.082920 4931 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.082939 4931 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-console-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.082958 4931 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-service-ca\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.082975 4931 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.509201 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vvmnt_2f63c980-6d60-4de2-8b1a-6a3e5d1b3864/console/0.log" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.509285 4931 generic.go:334] "Generic (PLEG): container finished" podID="2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" containerID="160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd" exitCode=2 Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.509443 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vvmnt" event={"ID":"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864","Type":"ContainerDied","Data":"160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd"} Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.509471 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vvmnt" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.509490 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vvmnt" event={"ID":"2f63c980-6d60-4de2-8b1a-6a3e5d1b3864","Type":"ContainerDied","Data":"405fb4605ca1de717c05ff6fa684bcff2ce130e3dc15a3a0fd5f4e31475d69ff"} Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.509520 4931 scope.go:117] "RemoveContainer" containerID="160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.513018 4931 generic.go:334] "Generic (PLEG): container finished" podID="86b012aa-97b4-461d-abf9-19a1a261b806" containerID="d88c3b4146b8b2fa15f16d00986e15da43b125e8362e30c589281feb6af7a355" exitCode=0 Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.513075 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" event={"ID":"86b012aa-97b4-461d-abf9-19a1a261b806","Type":"ContainerDied","Data":"d88c3b4146b8b2fa15f16d00986e15da43b125e8362e30c589281feb6af7a355"} Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.513147 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" event={"ID":"86b012aa-97b4-461d-abf9-19a1a261b806","Type":"ContainerStarted","Data":"c6ba17ba3fb432e150d6e1b760f2f0a28795b6b4280db231a86cce95603dba17"} Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.547963 4931 scope.go:117] "RemoveContainer" containerID="160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd" Nov 29 00:13:20 crc kubenswrapper[4931]: E1129 00:13:20.548987 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd\": container with ID starting with 160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd not found: ID does not exist" containerID="160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.549047 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd"} err="failed to get container status \"160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd\": rpc error: code = NotFound desc = could not find container \"160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd\": container with ID starting with 160018a9d4f94a7cccf0b57ef71968b72c667bb52761cc92c0f33ad8a5dae2dd not found: ID does not exist" Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.578798 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vvmnt"] Nov 29 00:13:20 crc kubenswrapper[4931]: I1129 00:13:20.588189 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-vvmnt"] Nov 29 00:13:21 crc kubenswrapper[4931]: I1129 00:13:21.226539 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" path="/var/lib/kubelet/pods/2f63c980-6d60-4de2-8b1a-6a3e5d1b3864/volumes" Nov 29 00:13:22 crc kubenswrapper[4931]: I1129 00:13:22.528290 4931 generic.go:334] "Generic (PLEG): container finished" podID="86b012aa-97b4-461d-abf9-19a1a261b806" containerID="7251a0111daaabf53987c61de82c2ce5c330ae8cd53f79dbce35bf91466317a9" exitCode=0 Nov 29 00:13:22 crc kubenswrapper[4931]: I1129 00:13:22.528391 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" event={"ID":"86b012aa-97b4-461d-abf9-19a1a261b806","Type":"ContainerDied","Data":"7251a0111daaabf53987c61de82c2ce5c330ae8cd53f79dbce35bf91466317a9"} Nov 29 00:13:23 crc kubenswrapper[4931]: I1129 00:13:23.539144 4931 generic.go:334] "Generic (PLEG): container finished" podID="86b012aa-97b4-461d-abf9-19a1a261b806" containerID="60c7a6975a3c698a7f100781c9cae58d654e4bf53a75fc0990a12ad6699a7e77" exitCode=0 Nov 29 00:13:23 crc kubenswrapper[4931]: I1129 00:13:23.539386 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" event={"ID":"86b012aa-97b4-461d-abf9-19a1a261b806","Type":"ContainerDied","Data":"60c7a6975a3c698a7f100781c9cae58d654e4bf53a75fc0990a12ad6699a7e77"} Nov 29 00:13:23 crc kubenswrapper[4931]: I1129 00:13:23.736007 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:13:23 crc kubenswrapper[4931]: I1129 00:13:23.736083 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:13:24 crc kubenswrapper[4931]: I1129 00:13:24.786281 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:24 crc kubenswrapper[4931]: I1129 00:13:24.946006 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86b012aa-97b4-461d-abf9-19a1a261b806-bundle\") pod \"86b012aa-97b4-461d-abf9-19a1a261b806\" (UID: \"86b012aa-97b4-461d-abf9-19a1a261b806\") " Nov 29 00:13:24 crc kubenswrapper[4931]: I1129 00:13:24.946140 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6ffk\" (UniqueName: \"kubernetes.io/projected/86b012aa-97b4-461d-abf9-19a1a261b806-kube-api-access-b6ffk\") pod \"86b012aa-97b4-461d-abf9-19a1a261b806\" (UID: \"86b012aa-97b4-461d-abf9-19a1a261b806\") " Nov 29 00:13:24 crc kubenswrapper[4931]: I1129 00:13:24.946175 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86b012aa-97b4-461d-abf9-19a1a261b806-util\") pod \"86b012aa-97b4-461d-abf9-19a1a261b806\" (UID: \"86b012aa-97b4-461d-abf9-19a1a261b806\") " Nov 29 00:13:24 crc kubenswrapper[4931]: I1129 00:13:24.947729 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86b012aa-97b4-461d-abf9-19a1a261b806-bundle" (OuterVolumeSpecName: "bundle") pod "86b012aa-97b4-461d-abf9-19a1a261b806" (UID: "86b012aa-97b4-461d-abf9-19a1a261b806"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:13:24 crc kubenswrapper[4931]: I1129 00:13:24.952695 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86b012aa-97b4-461d-abf9-19a1a261b806-kube-api-access-b6ffk" (OuterVolumeSpecName: "kube-api-access-b6ffk") pod "86b012aa-97b4-461d-abf9-19a1a261b806" (UID: "86b012aa-97b4-461d-abf9-19a1a261b806"). InnerVolumeSpecName "kube-api-access-b6ffk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:13:24 crc kubenswrapper[4931]: I1129 00:13:24.976984 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86b012aa-97b4-461d-abf9-19a1a261b806-util" (OuterVolumeSpecName: "util") pod "86b012aa-97b4-461d-abf9-19a1a261b806" (UID: "86b012aa-97b4-461d-abf9-19a1a261b806"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:13:25 crc kubenswrapper[4931]: I1129 00:13:25.049367 4931 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86b012aa-97b4-461d-abf9-19a1a261b806-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:25 crc kubenswrapper[4931]: I1129 00:13:25.049427 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6ffk\" (UniqueName: \"kubernetes.io/projected/86b012aa-97b4-461d-abf9-19a1a261b806-kube-api-access-b6ffk\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:25 crc kubenswrapper[4931]: I1129 00:13:25.049441 4931 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86b012aa-97b4-461d-abf9-19a1a261b806-util\") on node \"crc\" DevicePath \"\"" Nov 29 00:13:25 crc kubenswrapper[4931]: I1129 00:13:25.556214 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" event={"ID":"86b012aa-97b4-461d-abf9-19a1a261b806","Type":"ContainerDied","Data":"c6ba17ba3fb432e150d6e1b760f2f0a28795b6b4280db231a86cce95603dba17"} Nov 29 00:13:25 crc kubenswrapper[4931]: I1129 00:13:25.556255 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6ba17ba3fb432e150d6e1b760f2f0a28795b6b4280db231a86cce95603dba17" Nov 29 00:13:25 crc kubenswrapper[4931]: I1129 00:13:25.556606 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.471081 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72"] Nov 29 00:13:34 crc kubenswrapper[4931]: E1129 00:13:34.472171 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b012aa-97b4-461d-abf9-19a1a261b806" containerName="util" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.472198 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b012aa-97b4-461d-abf9-19a1a261b806" containerName="util" Nov 29 00:13:34 crc kubenswrapper[4931]: E1129 00:13:34.472234 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b012aa-97b4-461d-abf9-19a1a261b806" containerName="pull" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.472250 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b012aa-97b4-461d-abf9-19a1a261b806" containerName="pull" Nov 29 00:13:34 crc kubenswrapper[4931]: E1129 00:13:34.472285 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" containerName="console" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.472302 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" containerName="console" Nov 29 00:13:34 crc kubenswrapper[4931]: E1129 00:13:34.472327 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b012aa-97b4-461d-abf9-19a1a261b806" containerName="extract" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.472343 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b012aa-97b4-461d-abf9-19a1a261b806" containerName="extract" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.472572 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f63c980-6d60-4de2-8b1a-6a3e5d1b3864" containerName="console" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.472599 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="86b012aa-97b4-461d-abf9-19a1a261b806" containerName="extract" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.473282 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.476091 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.476605 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.476658 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-fscdn" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.476674 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.476680 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.483639 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72"] Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.492074 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/70a74ef6-a92b-49b3-b533-17b6a15f513c-webhook-cert\") pod \"metallb-operator-controller-manager-df4f9f548-5vp72\" (UID: \"70a74ef6-a92b-49b3-b533-17b6a15f513c\") " pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.492145 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnz72\" (UniqueName: \"kubernetes.io/projected/70a74ef6-a92b-49b3-b533-17b6a15f513c-kube-api-access-hnz72\") pod \"metallb-operator-controller-manager-df4f9f548-5vp72\" (UID: \"70a74ef6-a92b-49b3-b533-17b6a15f513c\") " pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.492180 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/70a74ef6-a92b-49b3-b533-17b6a15f513c-apiservice-cert\") pod \"metallb-operator-controller-manager-df4f9f548-5vp72\" (UID: \"70a74ef6-a92b-49b3-b533-17b6a15f513c\") " pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.592966 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/70a74ef6-a92b-49b3-b533-17b6a15f513c-webhook-cert\") pod \"metallb-operator-controller-manager-df4f9f548-5vp72\" (UID: \"70a74ef6-a92b-49b3-b533-17b6a15f513c\") " pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.593039 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnz72\" (UniqueName: \"kubernetes.io/projected/70a74ef6-a92b-49b3-b533-17b6a15f513c-kube-api-access-hnz72\") pod \"metallb-operator-controller-manager-df4f9f548-5vp72\" (UID: \"70a74ef6-a92b-49b3-b533-17b6a15f513c\") " pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.593072 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/70a74ef6-a92b-49b3-b533-17b6a15f513c-apiservice-cert\") pod \"metallb-operator-controller-manager-df4f9f548-5vp72\" (UID: \"70a74ef6-a92b-49b3-b533-17b6a15f513c\") " pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.599459 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/70a74ef6-a92b-49b3-b533-17b6a15f513c-webhook-cert\") pod \"metallb-operator-controller-manager-df4f9f548-5vp72\" (UID: \"70a74ef6-a92b-49b3-b533-17b6a15f513c\") " pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.599592 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/70a74ef6-a92b-49b3-b533-17b6a15f513c-apiservice-cert\") pod \"metallb-operator-controller-manager-df4f9f548-5vp72\" (UID: \"70a74ef6-a92b-49b3-b533-17b6a15f513c\") " pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.622860 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnz72\" (UniqueName: \"kubernetes.io/projected/70a74ef6-a92b-49b3-b533-17b6a15f513c-kube-api-access-hnz72\") pod \"metallb-operator-controller-manager-df4f9f548-5vp72\" (UID: \"70a74ef6-a92b-49b3-b533-17b6a15f513c\") " pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.701179 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72"] Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.702431 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.707341 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.709899 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.709883 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-kfhx7" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.758420 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72"] Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.790020 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.795604 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/04d440db-7f7e-46fa-9654-58b389ea3ad7-webhook-cert\") pod \"metallb-operator-webhook-server-d7b5f7bd5-t4z72\" (UID: \"04d440db-7f7e-46fa-9654-58b389ea3ad7\") " pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.795643 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/04d440db-7f7e-46fa-9654-58b389ea3ad7-apiservice-cert\") pod \"metallb-operator-webhook-server-d7b5f7bd5-t4z72\" (UID: \"04d440db-7f7e-46fa-9654-58b389ea3ad7\") " pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.795664 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ckw4\" (UniqueName: \"kubernetes.io/projected/04d440db-7f7e-46fa-9654-58b389ea3ad7-kube-api-access-8ckw4\") pod \"metallb-operator-webhook-server-d7b5f7bd5-t4z72\" (UID: \"04d440db-7f7e-46fa-9654-58b389ea3ad7\") " pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.897170 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/04d440db-7f7e-46fa-9654-58b389ea3ad7-webhook-cert\") pod \"metallb-operator-webhook-server-d7b5f7bd5-t4z72\" (UID: \"04d440db-7f7e-46fa-9654-58b389ea3ad7\") " pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.897440 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/04d440db-7f7e-46fa-9654-58b389ea3ad7-apiservice-cert\") pod \"metallb-operator-webhook-server-d7b5f7bd5-t4z72\" (UID: \"04d440db-7f7e-46fa-9654-58b389ea3ad7\") " pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.897919 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ckw4\" (UniqueName: \"kubernetes.io/projected/04d440db-7f7e-46fa-9654-58b389ea3ad7-kube-api-access-8ckw4\") pod \"metallb-operator-webhook-server-d7b5f7bd5-t4z72\" (UID: \"04d440db-7f7e-46fa-9654-58b389ea3ad7\") " pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.902735 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/04d440db-7f7e-46fa-9654-58b389ea3ad7-webhook-cert\") pod \"metallb-operator-webhook-server-d7b5f7bd5-t4z72\" (UID: \"04d440db-7f7e-46fa-9654-58b389ea3ad7\") " pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.910687 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/04d440db-7f7e-46fa-9654-58b389ea3ad7-apiservice-cert\") pod \"metallb-operator-webhook-server-d7b5f7bd5-t4z72\" (UID: \"04d440db-7f7e-46fa-9654-58b389ea3ad7\") " pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:13:34 crc kubenswrapper[4931]: I1129 00:13:34.920124 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ckw4\" (UniqueName: \"kubernetes.io/projected/04d440db-7f7e-46fa-9654-58b389ea3ad7-kube-api-access-8ckw4\") pod \"metallb-operator-webhook-server-d7b5f7bd5-t4z72\" (UID: \"04d440db-7f7e-46fa-9654-58b389ea3ad7\") " pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:13:35 crc kubenswrapper[4931]: I1129 00:13:35.038339 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:13:35 crc kubenswrapper[4931]: I1129 00:13:35.040393 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72"] Nov 29 00:13:35 crc kubenswrapper[4931]: I1129 00:13:35.490014 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72"] Nov 29 00:13:35 crc kubenswrapper[4931]: W1129 00:13:35.502757 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04d440db_7f7e_46fa_9654_58b389ea3ad7.slice/crio-9e69df32eae269232b711f76932d4d59d9346f37353f1e7b43b8d3bb576f5e37 WatchSource:0}: Error finding container 9e69df32eae269232b711f76932d4d59d9346f37353f1e7b43b8d3bb576f5e37: Status 404 returned error can't find the container with id 9e69df32eae269232b711f76932d4d59d9346f37353f1e7b43b8d3bb576f5e37 Nov 29 00:13:35 crc kubenswrapper[4931]: I1129 00:13:35.612434 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" event={"ID":"70a74ef6-a92b-49b3-b533-17b6a15f513c","Type":"ContainerStarted","Data":"670ae6ebbac8bc2d49cbb31c61c0015a2d46c98220d9084dee8b5652d8398677"} Nov 29 00:13:35 crc kubenswrapper[4931]: I1129 00:13:35.613350 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" event={"ID":"04d440db-7f7e-46fa-9654-58b389ea3ad7","Type":"ContainerStarted","Data":"9e69df32eae269232b711f76932d4d59d9346f37353f1e7b43b8d3bb576f5e37"} Nov 29 00:13:40 crc kubenswrapper[4931]: I1129 00:13:40.642417 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" event={"ID":"70a74ef6-a92b-49b3-b533-17b6a15f513c","Type":"ContainerStarted","Data":"b4aa28dd3742248e9c384c2a5e754fe3b0c3a6fa635f7196643e866e3436569b"} Nov 29 00:13:40 crc kubenswrapper[4931]: I1129 00:13:40.643215 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:13:40 crc kubenswrapper[4931]: I1129 00:13:40.644693 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" event={"ID":"04d440db-7f7e-46fa-9654-58b389ea3ad7","Type":"ContainerStarted","Data":"3e11086b691cf1f2dbaea3cddfd7f1fd4e7d2bf219060d80d43ecf75ee84e667"} Nov 29 00:13:40 crc kubenswrapper[4931]: I1129 00:13:40.645342 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:13:40 crc kubenswrapper[4931]: I1129 00:13:40.677003 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" podStartSLOduration=1.860849205 podStartE2EDuration="6.67697495s" podCreationTimestamp="2025-11-29 00:13:34 +0000 UTC" firstStartedPulling="2025-11-29 00:13:35.052043074 +0000 UTC m=+828.213936306" lastFinishedPulling="2025-11-29 00:13:39.868168819 +0000 UTC m=+833.030062051" observedRunningTime="2025-11-29 00:13:40.667072861 +0000 UTC m=+833.828966093" watchObservedRunningTime="2025-11-29 00:13:40.67697495 +0000 UTC m=+833.838868212" Nov 29 00:13:40 crc kubenswrapper[4931]: I1129 00:13:40.700178 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" podStartSLOduration=2.327381434 podStartE2EDuration="6.700161508s" podCreationTimestamp="2025-11-29 00:13:34 +0000 UTC" firstStartedPulling="2025-11-29 00:13:35.509805179 +0000 UTC m=+828.671698411" lastFinishedPulling="2025-11-29 00:13:39.882585253 +0000 UTC m=+833.044478485" observedRunningTime="2025-11-29 00:13:40.694725789 +0000 UTC m=+833.856619031" watchObservedRunningTime="2025-11-29 00:13:40.700161508 +0000 UTC m=+833.862054740" Nov 29 00:13:53 crc kubenswrapper[4931]: I1129 00:13:53.734870 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:13:53 crc kubenswrapper[4931]: I1129 00:13:53.735369 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:13:53 crc kubenswrapper[4931]: I1129 00:13:53.735410 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:13:53 crc kubenswrapper[4931]: I1129 00:13:53.735935 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b74c2da42faa59386db14c3bb51aa4674c697dd1c4383118b269448576e97f17"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 00:13:53 crc kubenswrapper[4931]: I1129 00:13:53.735979 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://b74c2da42faa59386db14c3bb51aa4674c697dd1c4383118b269448576e97f17" gracePeriod=600 Nov 29 00:13:54 crc kubenswrapper[4931]: I1129 00:13:54.745388 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="b74c2da42faa59386db14c3bb51aa4674c697dd1c4383118b269448576e97f17" exitCode=0 Nov 29 00:13:54 crc kubenswrapper[4931]: I1129 00:13:54.745461 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"b74c2da42faa59386db14c3bb51aa4674c697dd1c4383118b269448576e97f17"} Nov 29 00:13:54 crc kubenswrapper[4931]: I1129 00:13:54.746675 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"63e24d3607eb4b6f4826de9209c8d28131ebf746bafba6dd38a837ce70168944"} Nov 29 00:13:54 crc kubenswrapper[4931]: I1129 00:13:54.746719 4931 scope.go:117] "RemoveContainer" containerID="ffd259621969111fa5cf386e3c4f80aaf7be83cdeb850b46d5cecccd3d70c94e" Nov 29 00:13:55 crc kubenswrapper[4931]: I1129 00:13:55.043056 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-d7b5f7bd5-t4z72" Nov 29 00:14:14 crc kubenswrapper[4931]: I1129 00:14:14.793992 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-df4f9f548-5vp72" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.535512 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv"] Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.536453 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.539540 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-kwrbx" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.539756 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.543605 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-dff5k"] Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.546396 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.555237 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.555278 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv"] Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.555246 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.571291 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp4hd\" (UniqueName: \"kubernetes.io/projected/fe08c847-a4ed-4e65-a3cb-e2145c69365f-kube-api-access-gp4hd\") pod \"frr-k8s-webhook-server-7fcb986d4-c96lv\" (UID: \"fe08c847-a4ed-4e65-a3cb-e2145c69365f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.571369 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fe08c847-a4ed-4e65-a3cb-e2145c69365f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c96lv\" (UID: \"fe08c847-a4ed-4e65-a3cb-e2145c69365f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.641478 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-6ck62"] Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.642442 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6ck62" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.646207 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.646381 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.646423 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.646502 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-8gfxv" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.667172 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-qzb8h"] Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.668046 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.669457 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672063 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2eb6f229-405c-40db-8504-8fb851cb98d4-metallb-excludel2\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672103 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-frr-sockets\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672138 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gp4hd\" (UniqueName: \"kubernetes.io/projected/fe08c847-a4ed-4e65-a3cb-e2145c69365f-kube-api-access-gp4hd\") pod \"frr-k8s-webhook-server-7fcb986d4-c96lv\" (UID: \"fe08c847-a4ed-4e65-a3cb-e2145c69365f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672162 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dm5s\" (UniqueName: \"kubernetes.io/projected/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-kube-api-access-4dm5s\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672181 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-reloader\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672205 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-metrics\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672221 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fe08c847-a4ed-4e65-a3cb-e2145c69365f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c96lv\" (UID: \"fe08c847-a4ed-4e65-a3cb-e2145c69365f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672235 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p59mx\" (UniqueName: \"kubernetes.io/projected/2eb6f229-405c-40db-8504-8fb851cb98d4-kube-api-access-p59mx\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672251 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-metrics-certs\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672266 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-metrics-certs\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672279 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-frr-conf\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672304 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-memberlist\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.672320 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-frr-startup\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: E1129 00:14:15.672599 4931 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 29 00:14:15 crc kubenswrapper[4931]: E1129 00:14:15.672649 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe08c847-a4ed-4e65-a3cb-e2145c69365f-cert podName:fe08c847-a4ed-4e65-a3cb-e2145c69365f nodeName:}" failed. No retries permitted until 2025-11-29 00:14:16.172631779 +0000 UTC m=+869.334525181 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fe08c847-a4ed-4e65-a3cb-e2145c69365f-cert") pod "frr-k8s-webhook-server-7fcb986d4-c96lv" (UID: "fe08c847-a4ed-4e65-a3cb-e2145c69365f") : secret "frr-k8s-webhook-server-cert" not found Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.682423 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-qzb8h"] Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.693681 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gp4hd\" (UniqueName: \"kubernetes.io/projected/fe08c847-a4ed-4e65-a3cb-e2145c69365f-kube-api-access-gp4hd\") pod \"frr-k8s-webhook-server-7fcb986d4-c96lv\" (UID: \"fe08c847-a4ed-4e65-a3cb-e2145c69365f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773246 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-metrics\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773297 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p59mx\" (UniqueName: \"kubernetes.io/projected/2eb6f229-405c-40db-8504-8fb851cb98d4-kube-api-access-p59mx\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773338 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-metrics-certs\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773364 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-frr-conf\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773382 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-metrics-certs\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773404 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a44839c5-bede-4c88-b3ff-75f0d06074a6-cert\") pod \"controller-f8648f98b-qzb8h\" (UID: \"a44839c5-bede-4c88-b3ff-75f0d06074a6\") " pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773419 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lq5d\" (UniqueName: \"kubernetes.io/projected/a44839c5-bede-4c88-b3ff-75f0d06074a6-kube-api-access-6lq5d\") pod \"controller-f8648f98b-qzb8h\" (UID: \"a44839c5-bede-4c88-b3ff-75f0d06074a6\") " pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773445 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-memberlist\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773471 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-frr-startup\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773500 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2eb6f229-405c-40db-8504-8fb851cb98d4-metallb-excludel2\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773526 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-frr-sockets\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773554 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a44839c5-bede-4c88-b3ff-75f0d06074a6-metrics-certs\") pod \"controller-f8648f98b-qzb8h\" (UID: \"a44839c5-bede-4c88-b3ff-75f0d06074a6\") " pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773598 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dm5s\" (UniqueName: \"kubernetes.io/projected/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-kube-api-access-4dm5s\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: E1129 00:14:15.773618 4931 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 29 00:14:15 crc kubenswrapper[4931]: E1129 00:14:15.773673 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-metrics-certs podName:2eb6f229-405c-40db-8504-8fb851cb98d4 nodeName:}" failed. No retries permitted until 2025-11-29 00:14:16.27365928 +0000 UTC m=+869.435552512 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-metrics-certs") pod "speaker-6ck62" (UID: "2eb6f229-405c-40db-8504-8fb851cb98d4") : secret "speaker-certs-secret" not found Nov 29 00:14:15 crc kubenswrapper[4931]: E1129 00:14:15.773690 4931 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 29 00:14:15 crc kubenswrapper[4931]: E1129 00:14:15.773755 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-memberlist podName:2eb6f229-405c-40db-8504-8fb851cb98d4 nodeName:}" failed. No retries permitted until 2025-11-29 00:14:16.273730892 +0000 UTC m=+869.435624184 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-memberlist") pod "speaker-6ck62" (UID: "2eb6f229-405c-40db-8504-8fb851cb98d4") : secret "metallb-memberlist" not found Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773622 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-reloader\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.773969 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-metrics\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.774118 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-reloader\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.774318 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-frr-sockets\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.774536 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-frr-conf\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.774639 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2eb6f229-405c-40db-8504-8fb851cb98d4-metallb-excludel2\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.774901 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-frr-startup\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.778089 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-metrics-certs\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.807648 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p59mx\" (UniqueName: \"kubernetes.io/projected/2eb6f229-405c-40db-8504-8fb851cb98d4-kube-api-access-p59mx\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.826474 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dm5s\" (UniqueName: \"kubernetes.io/projected/65f9ee6d-ee55-45a6-abaf-22b0c059aad6-kube-api-access-4dm5s\") pod \"frr-k8s-dff5k\" (UID: \"65f9ee6d-ee55-45a6-abaf-22b0c059aad6\") " pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.869094 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.876629 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a44839c5-bede-4c88-b3ff-75f0d06074a6-metrics-certs\") pod \"controller-f8648f98b-qzb8h\" (UID: \"a44839c5-bede-4c88-b3ff-75f0d06074a6\") " pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.876718 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a44839c5-bede-4c88-b3ff-75f0d06074a6-cert\") pod \"controller-f8648f98b-qzb8h\" (UID: \"a44839c5-bede-4c88-b3ff-75f0d06074a6\") " pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.876738 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lq5d\" (UniqueName: \"kubernetes.io/projected/a44839c5-bede-4c88-b3ff-75f0d06074a6-kube-api-access-6lq5d\") pod \"controller-f8648f98b-qzb8h\" (UID: \"a44839c5-bede-4c88-b3ff-75f0d06074a6\") " pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.879229 4931 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.881982 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a44839c5-bede-4c88-b3ff-75f0d06074a6-metrics-certs\") pod \"controller-f8648f98b-qzb8h\" (UID: \"a44839c5-bede-4c88-b3ff-75f0d06074a6\") " pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.891331 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a44839c5-bede-4c88-b3ff-75f0d06074a6-cert\") pod \"controller-f8648f98b-qzb8h\" (UID: \"a44839c5-bede-4c88-b3ff-75f0d06074a6\") " pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.897732 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lq5d\" (UniqueName: \"kubernetes.io/projected/a44839c5-bede-4c88-b3ff-75f0d06074a6-kube-api-access-6lq5d\") pod \"controller-f8648f98b-qzb8h\" (UID: \"a44839c5-bede-4c88-b3ff-75f0d06074a6\") " pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:15 crc kubenswrapper[4931]: I1129 00:14:15.979322 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.181434 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fe08c847-a4ed-4e65-a3cb-e2145c69365f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c96lv\" (UID: \"fe08c847-a4ed-4e65-a3cb-e2145c69365f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.184640 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fe08c847-a4ed-4e65-a3cb-e2145c69365f-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c96lv\" (UID: \"fe08c847-a4ed-4e65-a3cb-e2145c69365f\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.283207 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-metrics-certs\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.283285 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-memberlist\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:16 crc kubenswrapper[4931]: E1129 00:14:16.285217 4931 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 29 00:14:16 crc kubenswrapper[4931]: E1129 00:14:16.285340 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-memberlist podName:2eb6f229-405c-40db-8504-8fb851cb98d4 nodeName:}" failed. No retries permitted until 2025-11-29 00:14:17.285279855 +0000 UTC m=+870.447173127 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-memberlist") pod "speaker-6ck62" (UID: "2eb6f229-405c-40db-8504-8fb851cb98d4") : secret "metallb-memberlist" not found Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.288178 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-metrics-certs\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.442020 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-qzb8h"] Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.457372 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.891486 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv"] Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.897117 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-qzb8h" event={"ID":"a44839c5-bede-4c88-b3ff-75f0d06074a6","Type":"ContainerStarted","Data":"ecad9dc443cb1b0fcc643c5c7a383a4aae205bdece6683f17b28acd7cb4c1c95"} Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.897249 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-qzb8h" event={"ID":"a44839c5-bede-4c88-b3ff-75f0d06074a6","Type":"ContainerStarted","Data":"7b798083e18f55ffed4bbe1ae41bbd9cd1ba3b449c7feea0284d4a45702a7136"} Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.897343 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-qzb8h" event={"ID":"a44839c5-bede-4c88-b3ff-75f0d06074a6","Type":"ContainerStarted","Data":"b274e679c2be29f1efa52600d10d80e3f81a9aebe4476a66303de1cf00fa6262"} Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.898056 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.900149 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dff5k" event={"ID":"65f9ee6d-ee55-45a6-abaf-22b0c059aad6","Type":"ContainerStarted","Data":"21f08b64c1d46371d8feab6ddec579e1eeba98b016c467e2fb293311eda840a5"} Nov 29 00:14:16 crc kubenswrapper[4931]: W1129 00:14:16.900861 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe08c847_a4ed_4e65_a3cb_e2145c69365f.slice/crio-048c823c97e59ddbcf3435516bb4a39d088ae1f891350566de97c239331d3b93 WatchSource:0}: Error finding container 048c823c97e59ddbcf3435516bb4a39d088ae1f891350566de97c239331d3b93: Status 404 returned error can't find the container with id 048c823c97e59ddbcf3435516bb4a39d088ae1f891350566de97c239331d3b93 Nov 29 00:14:16 crc kubenswrapper[4931]: I1129 00:14:16.917103 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-qzb8h" podStartSLOduration=1.917060062 podStartE2EDuration="1.917060062s" podCreationTimestamp="2025-11-29 00:14:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:14:16.914836661 +0000 UTC m=+870.076729923" watchObservedRunningTime="2025-11-29 00:14:16.917060062 +0000 UTC m=+870.078953334" Nov 29 00:14:17 crc kubenswrapper[4931]: I1129 00:14:17.296557 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-memberlist\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:17 crc kubenswrapper[4931]: I1129 00:14:17.308553 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2eb6f229-405c-40db-8504-8fb851cb98d4-memberlist\") pod \"speaker-6ck62\" (UID: \"2eb6f229-405c-40db-8504-8fb851cb98d4\") " pod="metallb-system/speaker-6ck62" Nov 29 00:14:17 crc kubenswrapper[4931]: I1129 00:14:17.454748 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6ck62" Nov 29 00:14:17 crc kubenswrapper[4931]: W1129 00:14:17.495967 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2eb6f229_405c_40db_8504_8fb851cb98d4.slice/crio-68218228d7fdc60af60612de4c7e4ad85ec5ed3f2d5e0f02411fd6d5a455a390 WatchSource:0}: Error finding container 68218228d7fdc60af60612de4c7e4ad85ec5ed3f2d5e0f02411fd6d5a455a390: Status 404 returned error can't find the container with id 68218228d7fdc60af60612de4c7e4ad85ec5ed3f2d5e0f02411fd6d5a455a390 Nov 29 00:14:17 crc kubenswrapper[4931]: I1129 00:14:17.909697 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" event={"ID":"fe08c847-a4ed-4e65-a3cb-e2145c69365f","Type":"ContainerStarted","Data":"048c823c97e59ddbcf3435516bb4a39d088ae1f891350566de97c239331d3b93"} Nov 29 00:14:17 crc kubenswrapper[4931]: I1129 00:14:17.911646 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6ck62" event={"ID":"2eb6f229-405c-40db-8504-8fb851cb98d4","Type":"ContainerStarted","Data":"94f5deaf745d861e3a03f2918b707e0fdddf53e09a27090e47e5fb2315734bbe"} Nov 29 00:14:17 crc kubenswrapper[4931]: I1129 00:14:17.911716 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6ck62" event={"ID":"2eb6f229-405c-40db-8504-8fb851cb98d4","Type":"ContainerStarted","Data":"68218228d7fdc60af60612de4c7e4ad85ec5ed3f2d5e0f02411fd6d5a455a390"} Nov 29 00:14:18 crc kubenswrapper[4931]: I1129 00:14:18.921733 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6ck62" event={"ID":"2eb6f229-405c-40db-8504-8fb851cb98d4","Type":"ContainerStarted","Data":"6cb0c6550b6365e0436f51966f3ea5711b8b256493910c7bf79416e397165b7e"} Nov 29 00:14:18 crc kubenswrapper[4931]: I1129 00:14:18.923009 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-6ck62" Nov 29 00:14:22 crc kubenswrapper[4931]: I1129 00:14:22.855622 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-6ck62" podStartSLOduration=7.85559708 podStartE2EDuration="7.85559708s" podCreationTimestamp="2025-11-29 00:14:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:14:18.947710107 +0000 UTC m=+872.109603349" watchObservedRunningTime="2025-11-29 00:14:22.85559708 +0000 UTC m=+876.017490322" Nov 29 00:14:22 crc kubenswrapper[4931]: I1129 00:14:22.856797 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w4tbs"] Nov 29 00:14:22 crc kubenswrapper[4931]: I1129 00:14:22.857976 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:22 crc kubenswrapper[4931]: I1129 00:14:22.870278 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w4tbs"] Nov 29 00:14:22 crc kubenswrapper[4931]: I1129 00:14:22.979839 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a7d0732-18ba-4c38-9d61-2f9797b300d3-catalog-content\") pod \"certified-operators-w4tbs\" (UID: \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\") " pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:22 crc kubenswrapper[4931]: I1129 00:14:22.979923 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a7d0732-18ba-4c38-9d61-2f9797b300d3-utilities\") pod \"certified-operators-w4tbs\" (UID: \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\") " pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:22 crc kubenswrapper[4931]: I1129 00:14:22.980028 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xmhr\" (UniqueName: \"kubernetes.io/projected/4a7d0732-18ba-4c38-9d61-2f9797b300d3-kube-api-access-5xmhr\") pod \"certified-operators-w4tbs\" (UID: \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\") " pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:23 crc kubenswrapper[4931]: I1129 00:14:23.081372 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xmhr\" (UniqueName: \"kubernetes.io/projected/4a7d0732-18ba-4c38-9d61-2f9797b300d3-kube-api-access-5xmhr\") pod \"certified-operators-w4tbs\" (UID: \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\") " pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:23 crc kubenswrapper[4931]: I1129 00:14:23.081441 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a7d0732-18ba-4c38-9d61-2f9797b300d3-catalog-content\") pod \"certified-operators-w4tbs\" (UID: \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\") " pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:23 crc kubenswrapper[4931]: I1129 00:14:23.081461 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a7d0732-18ba-4c38-9d61-2f9797b300d3-utilities\") pod \"certified-operators-w4tbs\" (UID: \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\") " pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:23 crc kubenswrapper[4931]: I1129 00:14:23.082209 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a7d0732-18ba-4c38-9d61-2f9797b300d3-utilities\") pod \"certified-operators-w4tbs\" (UID: \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\") " pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:23 crc kubenswrapper[4931]: I1129 00:14:23.082423 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a7d0732-18ba-4c38-9d61-2f9797b300d3-catalog-content\") pod \"certified-operators-w4tbs\" (UID: \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\") " pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:23 crc kubenswrapper[4931]: I1129 00:14:23.105744 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xmhr\" (UniqueName: \"kubernetes.io/projected/4a7d0732-18ba-4c38-9d61-2f9797b300d3-kube-api-access-5xmhr\") pod \"certified-operators-w4tbs\" (UID: \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\") " pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:23 crc kubenswrapper[4931]: I1129 00:14:23.188729 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:25 crc kubenswrapper[4931]: I1129 00:14:25.255593 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w4tbs"] Nov 29 00:14:25 crc kubenswrapper[4931]: W1129 00:14:25.281096 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a7d0732_18ba_4c38_9d61_2f9797b300d3.slice/crio-171b39ad42602140cddb4e36a00e686ae93813cfa765e8c811c35e0c4d6b3398 WatchSource:0}: Error finding container 171b39ad42602140cddb4e36a00e686ae93813cfa765e8c811c35e0c4d6b3398: Status 404 returned error can't find the container with id 171b39ad42602140cddb4e36a00e686ae93813cfa765e8c811c35e0c4d6b3398 Nov 29 00:14:25 crc kubenswrapper[4931]: I1129 00:14:25.965914 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" event={"ID":"fe08c847-a4ed-4e65-a3cb-e2145c69365f","Type":"ContainerStarted","Data":"3162fdff342e40c1ab45e5f518ce845ffa45b06390a10f6254139c8f492b1797"} Nov 29 00:14:25 crc kubenswrapper[4931]: I1129 00:14:25.966030 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" Nov 29 00:14:25 crc kubenswrapper[4931]: I1129 00:14:25.967795 4931 generic.go:334] "Generic (PLEG): container finished" podID="4a7d0732-18ba-4c38-9d61-2f9797b300d3" containerID="04028b0bd5a0775332bfaae1599760a90f22a3a5a5f63f5873f04b838a086227" exitCode=0 Nov 29 00:14:25 crc kubenswrapper[4931]: I1129 00:14:25.967959 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4tbs" event={"ID":"4a7d0732-18ba-4c38-9d61-2f9797b300d3","Type":"ContainerDied","Data":"04028b0bd5a0775332bfaae1599760a90f22a3a5a5f63f5873f04b838a086227"} Nov 29 00:14:25 crc kubenswrapper[4931]: I1129 00:14:25.967999 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4tbs" event={"ID":"4a7d0732-18ba-4c38-9d61-2f9797b300d3","Type":"ContainerStarted","Data":"171b39ad42602140cddb4e36a00e686ae93813cfa765e8c811c35e0c4d6b3398"} Nov 29 00:14:25 crc kubenswrapper[4931]: I1129 00:14:25.970024 4931 generic.go:334] "Generic (PLEG): container finished" podID="65f9ee6d-ee55-45a6-abaf-22b0c059aad6" containerID="4d833efc8f62107140db059f076c1a6f236452c5c37e2ad73ff350a888cb9054" exitCode=0 Nov 29 00:14:25 crc kubenswrapper[4931]: I1129 00:14:25.970084 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dff5k" event={"ID":"65f9ee6d-ee55-45a6-abaf-22b0c059aad6","Type":"ContainerDied","Data":"4d833efc8f62107140db059f076c1a6f236452c5c37e2ad73ff350a888cb9054"} Nov 29 00:14:26 crc kubenswrapper[4931]: I1129 00:14:26.000123 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" podStartSLOduration=2.767890173 podStartE2EDuration="11.000097508s" podCreationTimestamp="2025-11-29 00:14:15 +0000 UTC" firstStartedPulling="2025-11-29 00:14:16.906858379 +0000 UTC m=+870.068751621" lastFinishedPulling="2025-11-29 00:14:25.139065724 +0000 UTC m=+878.300958956" observedRunningTime="2025-11-29 00:14:25.989691259 +0000 UTC m=+879.151584531" watchObservedRunningTime="2025-11-29 00:14:26.000097508 +0000 UTC m=+879.161990780" Nov 29 00:14:26 crc kubenswrapper[4931]: I1129 00:14:26.981303 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4tbs" event={"ID":"4a7d0732-18ba-4c38-9d61-2f9797b300d3","Type":"ContainerStarted","Data":"c90fcb5f597aff78f88a01fa394eb0f002e2718590bd9680a02e1715c0f0806e"} Nov 29 00:14:26 crc kubenswrapper[4931]: I1129 00:14:26.983868 4931 generic.go:334] "Generic (PLEG): container finished" podID="65f9ee6d-ee55-45a6-abaf-22b0c059aad6" containerID="072fea8518172e1d2e00d17bd5bc72c9dda20b037ea039dea1f4e9b293b035aa" exitCode=0 Nov 29 00:14:26 crc kubenswrapper[4931]: I1129 00:14:26.983938 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dff5k" event={"ID":"65f9ee6d-ee55-45a6-abaf-22b0c059aad6","Type":"ContainerDied","Data":"072fea8518172e1d2e00d17bd5bc72c9dda20b037ea039dea1f4e9b293b035aa"} Nov 29 00:14:27 crc kubenswrapper[4931]: I1129 00:14:27.460000 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-6ck62" Nov 29 00:14:27 crc kubenswrapper[4931]: I1129 00:14:27.994069 4931 generic.go:334] "Generic (PLEG): container finished" podID="4a7d0732-18ba-4c38-9d61-2f9797b300d3" containerID="c90fcb5f597aff78f88a01fa394eb0f002e2718590bd9680a02e1715c0f0806e" exitCode=0 Nov 29 00:14:27 crc kubenswrapper[4931]: I1129 00:14:27.994356 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4tbs" event={"ID":"4a7d0732-18ba-4c38-9d61-2f9797b300d3","Type":"ContainerDied","Data":"c90fcb5f597aff78f88a01fa394eb0f002e2718590bd9680a02e1715c0f0806e"} Nov 29 00:14:27 crc kubenswrapper[4931]: I1129 00:14:27.998877 4931 generic.go:334] "Generic (PLEG): container finished" podID="65f9ee6d-ee55-45a6-abaf-22b0c059aad6" containerID="4a194ee02324482e3fc768cbc3f89171cb4fbe228786ed3b9a36ecb946602255" exitCode=0 Nov 29 00:14:27 crc kubenswrapper[4931]: I1129 00:14:27.998923 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dff5k" event={"ID":"65f9ee6d-ee55-45a6-abaf-22b0c059aad6","Type":"ContainerDied","Data":"4a194ee02324482e3fc768cbc3f89171cb4fbe228786ed3b9a36ecb946602255"} Nov 29 00:14:29 crc kubenswrapper[4931]: I1129 00:14:29.011143 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dff5k" event={"ID":"65f9ee6d-ee55-45a6-abaf-22b0c059aad6","Type":"ContainerStarted","Data":"cfcc86baa9903ae60435c528b518e32ac06f467d94a6f7d3368ce950a7f337f9"} Nov 29 00:14:29 crc kubenswrapper[4931]: I1129 00:14:29.011516 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dff5k" event={"ID":"65f9ee6d-ee55-45a6-abaf-22b0c059aad6","Type":"ContainerStarted","Data":"705aac43e362c87f206a75235482ac97799e04926fa93bd63261404267c1065a"} Nov 29 00:14:29 crc kubenswrapper[4931]: I1129 00:14:29.011527 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dff5k" event={"ID":"65f9ee6d-ee55-45a6-abaf-22b0c059aad6","Type":"ContainerStarted","Data":"c09febf085e070ca0436a9c548461a1cb46ed36ead7883b4f5738e9852f4e0ea"} Nov 29 00:14:29 crc kubenswrapper[4931]: I1129 00:14:29.011537 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dff5k" event={"ID":"65f9ee6d-ee55-45a6-abaf-22b0c059aad6","Type":"ContainerStarted","Data":"486c922456fde021be09760c9904de868d21c55153c02f0984d0ce0600a298f1"} Nov 29 00:14:30 crc kubenswrapper[4931]: I1129 00:14:30.028616 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4tbs" event={"ID":"4a7d0732-18ba-4c38-9d61-2f9797b300d3","Type":"ContainerStarted","Data":"a814e53737e3dfe4a13c7166152a884d01f33d89dc77757d1d5fc6ee058061e9"} Nov 29 00:14:30 crc kubenswrapper[4931]: I1129 00:14:30.047956 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dff5k" event={"ID":"65f9ee6d-ee55-45a6-abaf-22b0c059aad6","Type":"ContainerStarted","Data":"8f1ca72704cb14874a0cb0ee477377d3de128af41ba5bfb2b6fe402e4d524d97"} Nov 29 00:14:30 crc kubenswrapper[4931]: I1129 00:14:30.048035 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dff5k" event={"ID":"65f9ee6d-ee55-45a6-abaf-22b0c059aad6","Type":"ContainerStarted","Data":"11bdbc6c85bdbf14fe55b4a0b84fc4ffa2b2e01544b288bbf11e4adaf2bfb7f4"} Nov 29 00:14:30 crc kubenswrapper[4931]: I1129 00:14:30.048386 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:30 crc kubenswrapper[4931]: I1129 00:14:30.058708 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w4tbs" podStartSLOduration=5.447071577 podStartE2EDuration="8.058678139s" podCreationTimestamp="2025-11-29 00:14:22 +0000 UTC" firstStartedPulling="2025-11-29 00:14:25.969745066 +0000 UTC m=+879.131638328" lastFinishedPulling="2025-11-29 00:14:28.581351668 +0000 UTC m=+881.743244890" observedRunningTime="2025-11-29 00:14:30.056039786 +0000 UTC m=+883.217933058" watchObservedRunningTime="2025-11-29 00:14:30.058678139 +0000 UTC m=+883.220571421" Nov 29 00:14:30 crc kubenswrapper[4931]: I1129 00:14:30.089932 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-dff5k" podStartSLOduration=5.904178332 podStartE2EDuration="15.089909635s" podCreationTimestamp="2025-11-29 00:14:15 +0000 UTC" firstStartedPulling="2025-11-29 00:14:15.979195188 +0000 UTC m=+869.141088420" lastFinishedPulling="2025-11-29 00:14:25.164926481 +0000 UTC m=+878.326819723" observedRunningTime="2025-11-29 00:14:30.086222783 +0000 UTC m=+883.248116115" watchObservedRunningTime="2025-11-29 00:14:30.089909635 +0000 UTC m=+883.251802907" Nov 29 00:14:30 crc kubenswrapper[4931]: I1129 00:14:30.870476 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:30 crc kubenswrapper[4931]: I1129 00:14:30.913498 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:33 crc kubenswrapper[4931]: I1129 00:14:33.189946 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:33 crc kubenswrapper[4931]: I1129 00:14:33.190378 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:33 crc kubenswrapper[4931]: I1129 00:14:33.233551 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:34 crc kubenswrapper[4931]: I1129 00:14:34.129614 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:34 crc kubenswrapper[4931]: I1129 00:14:34.636184 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-cmgc6"] Nov 29 00:14:34 crc kubenswrapper[4931]: I1129 00:14:34.637076 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-cmgc6" Nov 29 00:14:34 crc kubenswrapper[4931]: I1129 00:14:34.638865 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-b4q4f" Nov 29 00:14:34 crc kubenswrapper[4931]: I1129 00:14:34.641866 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 29 00:14:34 crc kubenswrapper[4931]: I1129 00:14:34.642267 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 29 00:14:34 crc kubenswrapper[4931]: I1129 00:14:34.650025 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-cmgc6"] Nov 29 00:14:34 crc kubenswrapper[4931]: I1129 00:14:34.734651 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcc65\" (UniqueName: \"kubernetes.io/projected/88684874-2303-4a5b-93d6-5ceb28e68ef2-kube-api-access-xcc65\") pod \"openstack-operator-index-cmgc6\" (UID: \"88684874-2303-4a5b-93d6-5ceb28e68ef2\") " pod="openstack-operators/openstack-operator-index-cmgc6" Nov 29 00:14:34 crc kubenswrapper[4931]: I1129 00:14:34.836597 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcc65\" (UniqueName: \"kubernetes.io/projected/88684874-2303-4a5b-93d6-5ceb28e68ef2-kube-api-access-xcc65\") pod \"openstack-operator-index-cmgc6\" (UID: \"88684874-2303-4a5b-93d6-5ceb28e68ef2\") " pod="openstack-operators/openstack-operator-index-cmgc6" Nov 29 00:14:34 crc kubenswrapper[4931]: I1129 00:14:34.872489 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcc65\" (UniqueName: \"kubernetes.io/projected/88684874-2303-4a5b-93d6-5ceb28e68ef2-kube-api-access-xcc65\") pod \"openstack-operator-index-cmgc6\" (UID: \"88684874-2303-4a5b-93d6-5ceb28e68ef2\") " pod="openstack-operators/openstack-operator-index-cmgc6" Nov 29 00:14:34 crc kubenswrapper[4931]: I1129 00:14:34.971089 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-cmgc6" Nov 29 00:14:35 crc kubenswrapper[4931]: I1129 00:14:35.405887 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-cmgc6"] Nov 29 00:14:35 crc kubenswrapper[4931]: W1129 00:14:35.411515 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88684874_2303_4a5b_93d6_5ceb28e68ef2.slice/crio-8244299578e2d8570d2066817d55a289ac65529fe6b07da6abac7f795ddd6505 WatchSource:0}: Error finding container 8244299578e2d8570d2066817d55a289ac65529fe6b07da6abac7f795ddd6505: Status 404 returned error can't find the container with id 8244299578e2d8570d2066817d55a289ac65529fe6b07da6abac7f795ddd6505 Nov 29 00:14:35 crc kubenswrapper[4931]: I1129 00:14:35.982688 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-qzb8h" Nov 29 00:14:36 crc kubenswrapper[4931]: I1129 00:14:36.083426 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-cmgc6" event={"ID":"88684874-2303-4a5b-93d6-5ceb28e68ef2","Type":"ContainerStarted","Data":"8244299578e2d8570d2066817d55a289ac65529fe6b07da6abac7f795ddd6505"} Nov 29 00:14:36 crc kubenswrapper[4931]: I1129 00:14:36.465359 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c96lv" Nov 29 00:14:37 crc kubenswrapper[4931]: I1129 00:14:37.428472 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w4tbs"] Nov 29 00:14:37 crc kubenswrapper[4931]: I1129 00:14:37.429071 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w4tbs" podUID="4a7d0732-18ba-4c38-9d61-2f9797b300d3" containerName="registry-server" containerID="cri-o://a814e53737e3dfe4a13c7166152a884d01f33d89dc77757d1d5fc6ee058061e9" gracePeriod=2 Nov 29 00:14:38 crc kubenswrapper[4931]: I1129 00:14:38.104031 4931 generic.go:334] "Generic (PLEG): container finished" podID="4a7d0732-18ba-4c38-9d61-2f9797b300d3" containerID="a814e53737e3dfe4a13c7166152a884d01f33d89dc77757d1d5fc6ee058061e9" exitCode=0 Nov 29 00:14:38 crc kubenswrapper[4931]: I1129 00:14:38.104085 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4tbs" event={"ID":"4a7d0732-18ba-4c38-9d61-2f9797b300d3","Type":"ContainerDied","Data":"a814e53737e3dfe4a13c7166152a884d01f33d89dc77757d1d5fc6ee058061e9"} Nov 29 00:14:38 crc kubenswrapper[4931]: I1129 00:14:38.474951 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:38 crc kubenswrapper[4931]: I1129 00:14:38.624153 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a7d0732-18ba-4c38-9d61-2f9797b300d3-utilities\") pod \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\" (UID: \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\") " Nov 29 00:14:38 crc kubenswrapper[4931]: I1129 00:14:38.624261 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a7d0732-18ba-4c38-9d61-2f9797b300d3-catalog-content\") pod \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\" (UID: \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\") " Nov 29 00:14:38 crc kubenswrapper[4931]: I1129 00:14:38.624325 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xmhr\" (UniqueName: \"kubernetes.io/projected/4a7d0732-18ba-4c38-9d61-2f9797b300d3-kube-api-access-5xmhr\") pod \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\" (UID: \"4a7d0732-18ba-4c38-9d61-2f9797b300d3\") " Nov 29 00:14:38 crc kubenswrapper[4931]: I1129 00:14:38.625637 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a7d0732-18ba-4c38-9d61-2f9797b300d3-utilities" (OuterVolumeSpecName: "utilities") pod "4a7d0732-18ba-4c38-9d61-2f9797b300d3" (UID: "4a7d0732-18ba-4c38-9d61-2f9797b300d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:14:38 crc kubenswrapper[4931]: I1129 00:14:38.642078 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a7d0732-18ba-4c38-9d61-2f9797b300d3-kube-api-access-5xmhr" (OuterVolumeSpecName: "kube-api-access-5xmhr") pod "4a7d0732-18ba-4c38-9d61-2f9797b300d3" (UID: "4a7d0732-18ba-4c38-9d61-2f9797b300d3"). InnerVolumeSpecName "kube-api-access-5xmhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:14:38 crc kubenswrapper[4931]: I1129 00:14:38.675238 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a7d0732-18ba-4c38-9d61-2f9797b300d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a7d0732-18ba-4c38-9d61-2f9797b300d3" (UID: "4a7d0732-18ba-4c38-9d61-2f9797b300d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:14:38 crc kubenswrapper[4931]: I1129 00:14:38.727275 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a7d0732-18ba-4c38-9d61-2f9797b300d3-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:14:38 crc kubenswrapper[4931]: I1129 00:14:38.727309 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a7d0732-18ba-4c38-9d61-2f9797b300d3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:14:38 crc kubenswrapper[4931]: I1129 00:14:38.727326 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xmhr\" (UniqueName: \"kubernetes.io/projected/4a7d0732-18ba-4c38-9d61-2f9797b300d3-kube-api-access-5xmhr\") on node \"crc\" DevicePath \"\"" Nov 29 00:14:39 crc kubenswrapper[4931]: I1129 00:14:39.112122 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4tbs" event={"ID":"4a7d0732-18ba-4c38-9d61-2f9797b300d3","Type":"ContainerDied","Data":"171b39ad42602140cddb4e36a00e686ae93813cfa765e8c811c35e0c4d6b3398"} Nov 29 00:14:39 crc kubenswrapper[4931]: I1129 00:14:39.112458 4931 scope.go:117] "RemoveContainer" containerID="a814e53737e3dfe4a13c7166152a884d01f33d89dc77757d1d5fc6ee058061e9" Nov 29 00:14:39 crc kubenswrapper[4931]: I1129 00:14:39.112306 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w4tbs" Nov 29 00:14:39 crc kubenswrapper[4931]: I1129 00:14:39.148045 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w4tbs"] Nov 29 00:14:39 crc kubenswrapper[4931]: I1129 00:14:39.152362 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w4tbs"] Nov 29 00:14:39 crc kubenswrapper[4931]: I1129 00:14:39.218873 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a7d0732-18ba-4c38-9d61-2f9797b300d3" path="/var/lib/kubelet/pods/4a7d0732-18ba-4c38-9d61-2f9797b300d3/volumes" Nov 29 00:14:39 crc kubenswrapper[4931]: I1129 00:14:39.373662 4931 scope.go:117] "RemoveContainer" containerID="c90fcb5f597aff78f88a01fa394eb0f002e2718590bd9680a02e1715c0f0806e" Nov 29 00:14:39 crc kubenswrapper[4931]: I1129 00:14:39.387851 4931 scope.go:117] "RemoveContainer" containerID="04028b0bd5a0775332bfaae1599760a90f22a3a5a5f63f5873f04b838a086227" Nov 29 00:14:40 crc kubenswrapper[4931]: I1129 00:14:40.120280 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-cmgc6" event={"ID":"88684874-2303-4a5b-93d6-5ceb28e68ef2","Type":"ContainerStarted","Data":"ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507"} Nov 29 00:14:40 crc kubenswrapper[4931]: I1129 00:14:40.141294 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-cmgc6" podStartSLOduration=2.115356751 podStartE2EDuration="6.141276418s" podCreationTimestamp="2025-11-29 00:14:34 +0000 UTC" firstStartedPulling="2025-11-29 00:14:35.41385515 +0000 UTC m=+888.575748382" lastFinishedPulling="2025-11-29 00:14:39.439774817 +0000 UTC m=+892.601668049" observedRunningTime="2025-11-29 00:14:40.136945778 +0000 UTC m=+893.298839010" watchObservedRunningTime="2025-11-29 00:14:40.141276418 +0000 UTC m=+893.303169650" Nov 29 00:14:41 crc kubenswrapper[4931]: I1129 00:14:41.233635 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-cmgc6"] Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.040267 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-h7scz"] Nov 29 00:14:42 crc kubenswrapper[4931]: E1129 00:14:42.040717 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a7d0732-18ba-4c38-9d61-2f9797b300d3" containerName="registry-server" Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.040777 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a7d0732-18ba-4c38-9d61-2f9797b300d3" containerName="registry-server" Nov 29 00:14:42 crc kubenswrapper[4931]: E1129 00:14:42.040801 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a7d0732-18ba-4c38-9d61-2f9797b300d3" containerName="extract-content" Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.040890 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a7d0732-18ba-4c38-9d61-2f9797b300d3" containerName="extract-content" Nov 29 00:14:42 crc kubenswrapper[4931]: E1129 00:14:42.040929 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a7d0732-18ba-4c38-9d61-2f9797b300d3" containerName="extract-utilities" Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.040946 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a7d0732-18ba-4c38-9d61-2f9797b300d3" containerName="extract-utilities" Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.041212 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a7d0732-18ba-4c38-9d61-2f9797b300d3" containerName="registry-server" Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.042161 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-h7scz" Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.069257 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-h7scz"] Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.133246 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-cmgc6" podUID="88684874-2303-4a5b-93d6-5ceb28e68ef2" containerName="registry-server" containerID="cri-o://ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507" gracePeriod=2 Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.185906 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t264z\" (UniqueName: \"kubernetes.io/projected/4ef2db4a-ba6d-47a1-8054-b61f854da2d0-kube-api-access-t264z\") pod \"openstack-operator-index-h7scz\" (UID: \"4ef2db4a-ba6d-47a1-8054-b61f854da2d0\") " pod="openstack-operators/openstack-operator-index-h7scz" Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.287663 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t264z\" (UniqueName: \"kubernetes.io/projected/4ef2db4a-ba6d-47a1-8054-b61f854da2d0-kube-api-access-t264z\") pod \"openstack-operator-index-h7scz\" (UID: \"4ef2db4a-ba6d-47a1-8054-b61f854da2d0\") " pod="openstack-operators/openstack-operator-index-h7scz" Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.312680 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t264z\" (UniqueName: \"kubernetes.io/projected/4ef2db4a-ba6d-47a1-8054-b61f854da2d0-kube-api-access-t264z\") pod \"openstack-operator-index-h7scz\" (UID: \"4ef2db4a-ba6d-47a1-8054-b61f854da2d0\") " pod="openstack-operators/openstack-operator-index-h7scz" Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.379092 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-h7scz" Nov 29 00:14:42 crc kubenswrapper[4931]: I1129 00:14:42.917444 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-h7scz"] Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.087059 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-cmgc6" Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.139238 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-h7scz" event={"ID":"4ef2db4a-ba6d-47a1-8054-b61f854da2d0","Type":"ContainerStarted","Data":"8af87fd2780d97508ccdcff42a8009fdda16ee1893f589b52f0644f71db74145"} Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.139282 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-h7scz" event={"ID":"4ef2db4a-ba6d-47a1-8054-b61f854da2d0","Type":"ContainerStarted","Data":"eed3e7cb85cd665f936d19a02715bfe01b963207b71b43b132113b4a3f960041"} Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.142294 4931 generic.go:334] "Generic (PLEG): container finished" podID="88684874-2303-4a5b-93d6-5ceb28e68ef2" containerID="ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507" exitCode=0 Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.142333 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-cmgc6" event={"ID":"88684874-2303-4a5b-93d6-5ceb28e68ef2","Type":"ContainerDied","Data":"ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507"} Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.142355 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-cmgc6" event={"ID":"88684874-2303-4a5b-93d6-5ceb28e68ef2","Type":"ContainerDied","Data":"8244299578e2d8570d2066817d55a289ac65529fe6b07da6abac7f795ddd6505"} Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.142363 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-cmgc6" Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.142379 4931 scope.go:117] "RemoveContainer" containerID="ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507" Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.154916 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-h7scz" podStartSLOduration=1.100594829 podStartE2EDuration="1.154897995s" podCreationTimestamp="2025-11-29 00:14:42 +0000 UTC" firstStartedPulling="2025-11-29 00:14:42.941011655 +0000 UTC m=+896.102904907" lastFinishedPulling="2025-11-29 00:14:42.995314841 +0000 UTC m=+896.157208073" observedRunningTime="2025-11-29 00:14:43.152378825 +0000 UTC m=+896.314272077" watchObservedRunningTime="2025-11-29 00:14:43.154897995 +0000 UTC m=+896.316791237" Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.159499 4931 scope.go:117] "RemoveContainer" containerID="ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507" Nov 29 00:14:43 crc kubenswrapper[4931]: E1129 00:14:43.160010 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507\": container with ID starting with ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507 not found: ID does not exist" containerID="ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507" Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.160049 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507"} err="failed to get container status \"ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507\": rpc error: code = NotFound desc = could not find container \"ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507\": container with ID starting with ec42403c9bb430fb84dcefec445bd8bd448b851c79171e28c71e24fad8cd0507 not found: ID does not exist" Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.205046 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcc65\" (UniqueName: \"kubernetes.io/projected/88684874-2303-4a5b-93d6-5ceb28e68ef2-kube-api-access-xcc65\") pod \"88684874-2303-4a5b-93d6-5ceb28e68ef2\" (UID: \"88684874-2303-4a5b-93d6-5ceb28e68ef2\") " Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.212224 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88684874-2303-4a5b-93d6-5ceb28e68ef2-kube-api-access-xcc65" (OuterVolumeSpecName: "kube-api-access-xcc65") pod "88684874-2303-4a5b-93d6-5ceb28e68ef2" (UID: "88684874-2303-4a5b-93d6-5ceb28e68ef2"). InnerVolumeSpecName "kube-api-access-xcc65". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.306878 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcc65\" (UniqueName: \"kubernetes.io/projected/88684874-2303-4a5b-93d6-5ceb28e68ef2-kube-api-access-xcc65\") on node \"crc\" DevicePath \"\"" Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.469437 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-cmgc6"] Nov 29 00:14:43 crc kubenswrapper[4931]: I1129 00:14:43.476697 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-cmgc6"] Nov 29 00:14:45 crc kubenswrapper[4931]: I1129 00:14:45.225373 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88684874-2303-4a5b-93d6-5ceb28e68ef2" path="/var/lib/kubelet/pods/88684874-2303-4a5b-93d6-5ceb28e68ef2/volumes" Nov 29 00:14:45 crc kubenswrapper[4931]: I1129 00:14:45.873949 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-dff5k" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.442567 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9xfht"] Nov 29 00:14:46 crc kubenswrapper[4931]: E1129 00:14:46.443051 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88684874-2303-4a5b-93d6-5ceb28e68ef2" containerName="registry-server" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.443077 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="88684874-2303-4a5b-93d6-5ceb28e68ef2" containerName="registry-server" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.443269 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="88684874-2303-4a5b-93d6-5ceb28e68ef2" containerName="registry-server" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.448038 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.464064 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9xfht"] Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.553477 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af0c698a-6e96-4837-80b1-de46e1c404f4-utilities\") pod \"community-operators-9xfht\" (UID: \"af0c698a-6e96-4837-80b1-de46e1c404f4\") " pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.553529 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af0c698a-6e96-4837-80b1-de46e1c404f4-catalog-content\") pod \"community-operators-9xfht\" (UID: \"af0c698a-6e96-4837-80b1-de46e1c404f4\") " pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.553557 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpn9j\" (UniqueName: \"kubernetes.io/projected/af0c698a-6e96-4837-80b1-de46e1c404f4-kube-api-access-qpn9j\") pod \"community-operators-9xfht\" (UID: \"af0c698a-6e96-4837-80b1-de46e1c404f4\") " pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.655274 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af0c698a-6e96-4837-80b1-de46e1c404f4-utilities\") pod \"community-operators-9xfht\" (UID: \"af0c698a-6e96-4837-80b1-de46e1c404f4\") " pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.655325 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af0c698a-6e96-4837-80b1-de46e1c404f4-catalog-content\") pod \"community-operators-9xfht\" (UID: \"af0c698a-6e96-4837-80b1-de46e1c404f4\") " pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.655346 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpn9j\" (UniqueName: \"kubernetes.io/projected/af0c698a-6e96-4837-80b1-de46e1c404f4-kube-api-access-qpn9j\") pod \"community-operators-9xfht\" (UID: \"af0c698a-6e96-4837-80b1-de46e1c404f4\") " pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.655772 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af0c698a-6e96-4837-80b1-de46e1c404f4-utilities\") pod \"community-operators-9xfht\" (UID: \"af0c698a-6e96-4837-80b1-de46e1c404f4\") " pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.655935 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af0c698a-6e96-4837-80b1-de46e1c404f4-catalog-content\") pod \"community-operators-9xfht\" (UID: \"af0c698a-6e96-4837-80b1-de46e1c404f4\") " pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.681177 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpn9j\" (UniqueName: \"kubernetes.io/projected/af0c698a-6e96-4837-80b1-de46e1c404f4-kube-api-access-qpn9j\") pod \"community-operators-9xfht\" (UID: \"af0c698a-6e96-4837-80b1-de46e1c404f4\") " pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:46 crc kubenswrapper[4931]: I1129 00:14:46.770094 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:47 crc kubenswrapper[4931]: I1129 00:14:47.231203 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9xfht"] Nov 29 00:14:48 crc kubenswrapper[4931]: I1129 00:14:48.186040 4931 generic.go:334] "Generic (PLEG): container finished" podID="af0c698a-6e96-4837-80b1-de46e1c404f4" containerID="93a93893b6453f2db7f9fbf9e978bf652ba05d94b2b46ebadf978a146aa3e8fa" exitCode=0 Nov 29 00:14:48 crc kubenswrapper[4931]: I1129 00:14:48.186183 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfht" event={"ID":"af0c698a-6e96-4837-80b1-de46e1c404f4","Type":"ContainerDied","Data":"93a93893b6453f2db7f9fbf9e978bf652ba05d94b2b46ebadf978a146aa3e8fa"} Nov 29 00:14:48 crc kubenswrapper[4931]: I1129 00:14:48.186419 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfht" event={"ID":"af0c698a-6e96-4837-80b1-de46e1c404f4","Type":"ContainerStarted","Data":"b2b294e0bd9e2cd15511a11e7e053442144eeb604a60d292b0a5e6065310c1db"} Nov 29 00:14:49 crc kubenswrapper[4931]: I1129 00:14:49.197976 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfht" event={"ID":"af0c698a-6e96-4837-80b1-de46e1c404f4","Type":"ContainerStarted","Data":"490077cd77ee808d0875c196c04a44e29bec3f6ba4d07251cca283f6f25180cf"} Nov 29 00:14:50 crc kubenswrapper[4931]: I1129 00:14:50.209472 4931 generic.go:334] "Generic (PLEG): container finished" podID="af0c698a-6e96-4837-80b1-de46e1c404f4" containerID="490077cd77ee808d0875c196c04a44e29bec3f6ba4d07251cca283f6f25180cf" exitCode=0 Nov 29 00:14:50 crc kubenswrapper[4931]: I1129 00:14:50.209545 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfht" event={"ID":"af0c698a-6e96-4837-80b1-de46e1c404f4","Type":"ContainerDied","Data":"490077cd77ee808d0875c196c04a44e29bec3f6ba4d07251cca283f6f25180cf"} Nov 29 00:14:51 crc kubenswrapper[4931]: I1129 00:14:51.225537 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfht" event={"ID":"af0c698a-6e96-4837-80b1-de46e1c404f4","Type":"ContainerStarted","Data":"ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821"} Nov 29 00:14:51 crc kubenswrapper[4931]: I1129 00:14:51.260030 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9xfht" podStartSLOduration=2.830452601 podStartE2EDuration="5.260005295s" podCreationTimestamp="2025-11-29 00:14:46 +0000 UTC" firstStartedPulling="2025-11-29 00:14:48.188606904 +0000 UTC m=+901.350500126" lastFinishedPulling="2025-11-29 00:14:50.618159588 +0000 UTC m=+903.780052820" observedRunningTime="2025-11-29 00:14:51.251920141 +0000 UTC m=+904.413813443" watchObservedRunningTime="2025-11-29 00:14:51.260005295 +0000 UTC m=+904.421898557" Nov 29 00:14:52 crc kubenswrapper[4931]: I1129 00:14:52.379265 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-h7scz" Nov 29 00:14:52 crc kubenswrapper[4931]: I1129 00:14:52.379367 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-h7scz" Nov 29 00:14:52 crc kubenswrapper[4931]: I1129 00:14:52.417098 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-h7scz" Nov 29 00:14:53 crc kubenswrapper[4931]: I1129 00:14:53.281578 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-h7scz" Nov 29 00:14:56 crc kubenswrapper[4931]: I1129 00:14:56.770645 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:56 crc kubenswrapper[4931]: I1129 00:14:56.771233 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:56 crc kubenswrapper[4931]: I1129 00:14:56.821527 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:56 crc kubenswrapper[4931]: I1129 00:14:56.886709 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w"] Nov 29 00:14:56 crc kubenswrapper[4931]: I1129 00:14:56.888094 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:14:56 crc kubenswrapper[4931]: I1129 00:14:56.890583 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lgs7d" Nov 29 00:14:56 crc kubenswrapper[4931]: I1129 00:14:56.901444 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w"] Nov 29 00:14:57 crc kubenswrapper[4931]: I1129 00:14:57.005321 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8227\" (UniqueName: \"kubernetes.io/projected/ef059f7c-88cc-442c-9f6e-5e7a5778a549-kube-api-access-c8227\") pod \"287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w\" (UID: \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\") " pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:14:57 crc kubenswrapper[4931]: I1129 00:14:57.005375 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef059f7c-88cc-442c-9f6e-5e7a5778a549-util\") pod \"287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w\" (UID: \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\") " pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:14:57 crc kubenswrapper[4931]: I1129 00:14:57.005416 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef059f7c-88cc-442c-9f6e-5e7a5778a549-bundle\") pod \"287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w\" (UID: \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\") " pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:14:57 crc kubenswrapper[4931]: I1129 00:14:57.106768 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8227\" (UniqueName: \"kubernetes.io/projected/ef059f7c-88cc-442c-9f6e-5e7a5778a549-kube-api-access-c8227\") pod \"287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w\" (UID: \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\") " pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:14:57 crc kubenswrapper[4931]: I1129 00:14:57.106854 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef059f7c-88cc-442c-9f6e-5e7a5778a549-util\") pod \"287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w\" (UID: \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\") " pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:14:57 crc kubenswrapper[4931]: I1129 00:14:57.106902 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef059f7c-88cc-442c-9f6e-5e7a5778a549-bundle\") pod \"287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w\" (UID: \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\") " pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:14:57 crc kubenswrapper[4931]: I1129 00:14:57.107559 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef059f7c-88cc-442c-9f6e-5e7a5778a549-bundle\") pod \"287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w\" (UID: \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\") " pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:14:57 crc kubenswrapper[4931]: I1129 00:14:57.107800 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef059f7c-88cc-442c-9f6e-5e7a5778a549-util\") pod \"287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w\" (UID: \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\") " pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:14:57 crc kubenswrapper[4931]: I1129 00:14:57.134236 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8227\" (UniqueName: \"kubernetes.io/projected/ef059f7c-88cc-442c-9f6e-5e7a5778a549-kube-api-access-c8227\") pod \"287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w\" (UID: \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\") " pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:14:57 crc kubenswrapper[4931]: I1129 00:14:57.207996 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:14:57 crc kubenswrapper[4931]: I1129 00:14:57.356888 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:14:57 crc kubenswrapper[4931]: I1129 00:14:57.654516 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w"] Nov 29 00:14:58 crc kubenswrapper[4931]: I1129 00:14:58.289011 4931 generic.go:334] "Generic (PLEG): container finished" podID="ef059f7c-88cc-442c-9f6e-5e7a5778a549" containerID="2d72179bc6b45d3cd6ce659f1cc08424d35cccf7c5b3c132f0785aa18e846133" exitCode=0 Nov 29 00:14:58 crc kubenswrapper[4931]: I1129 00:14:58.289101 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" event={"ID":"ef059f7c-88cc-442c-9f6e-5e7a5778a549","Type":"ContainerDied","Data":"2d72179bc6b45d3cd6ce659f1cc08424d35cccf7c5b3c132f0785aa18e846133"} Nov 29 00:14:58 crc kubenswrapper[4931]: I1129 00:14:58.289649 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" event={"ID":"ef059f7c-88cc-442c-9f6e-5e7a5778a549","Type":"ContainerStarted","Data":"bc4864a139e95c085f8930011846d22bd38f8c756a3868f3de84d67638bd94ef"} Nov 29 00:14:59 crc kubenswrapper[4931]: I1129 00:14:59.299465 4931 generic.go:334] "Generic (PLEG): container finished" podID="ef059f7c-88cc-442c-9f6e-5e7a5778a549" containerID="156f88efc6025a22e883d9d275be54b1e06ac5882296f99a5ae31532fb31d108" exitCode=0 Nov 29 00:14:59 crc kubenswrapper[4931]: I1129 00:14:59.299517 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" event={"ID":"ef059f7c-88cc-442c-9f6e-5e7a5778a549","Type":"ContainerDied","Data":"156f88efc6025a22e883d9d275be54b1e06ac5882296f99a5ae31532fb31d108"} Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.163657 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj"] Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.164860 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.167768 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.168761 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.178981 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj"] Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.246852 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-secret-volume\") pod \"collect-profiles-29406255-pz6sj\" (UID: \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.246902 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-config-volume\") pod \"collect-profiles-29406255-pz6sj\" (UID: \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.247152 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmm6l\" (UniqueName: \"kubernetes.io/projected/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-kube-api-access-wmm6l\") pod \"collect-profiles-29406255-pz6sj\" (UID: \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.307921 4931 generic.go:334] "Generic (PLEG): container finished" podID="ef059f7c-88cc-442c-9f6e-5e7a5778a549" containerID="78068625bbaa452b4822ac97caecddd2b946edb077ea2f748568180833f9c503" exitCode=0 Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.307963 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" event={"ID":"ef059f7c-88cc-442c-9f6e-5e7a5778a549","Type":"ContainerDied","Data":"78068625bbaa452b4822ac97caecddd2b946edb077ea2f748568180833f9c503"} Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.348019 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmm6l\" (UniqueName: \"kubernetes.io/projected/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-kube-api-access-wmm6l\") pod \"collect-profiles-29406255-pz6sj\" (UID: \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.348079 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-secret-volume\") pod \"collect-profiles-29406255-pz6sj\" (UID: \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.348112 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-config-volume\") pod \"collect-profiles-29406255-pz6sj\" (UID: \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.349111 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-config-volume\") pod \"collect-profiles-29406255-pz6sj\" (UID: \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.359848 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-secret-volume\") pod \"collect-profiles-29406255-pz6sj\" (UID: \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.366150 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmm6l\" (UniqueName: \"kubernetes.io/projected/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-kube-api-access-wmm6l\") pod \"collect-profiles-29406255-pz6sj\" (UID: \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.496375 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:00 crc kubenswrapper[4931]: I1129 00:15:00.726761 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj"] Nov 29 00:15:00 crc kubenswrapper[4931]: W1129 00:15:00.730689 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f0cfdd5_2bcd_4709_bef6_82a3a1939e2a.slice/crio-596770696f28913fd59c02eefb975647c58f84ae718a66e65ec52d07e5904dc8 WatchSource:0}: Error finding container 596770696f28913fd59c02eefb975647c58f84ae718a66e65ec52d07e5904dc8: Status 404 returned error can't find the container with id 596770696f28913fd59c02eefb975647c58f84ae718a66e65ec52d07e5904dc8 Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.233037 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9xfht"] Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.233890 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9xfht" podUID="af0c698a-6e96-4837-80b1-de46e1c404f4" containerName="registry-server" containerID="cri-o://ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821" gracePeriod=2 Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.315996 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" event={"ID":"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a","Type":"ContainerStarted","Data":"596770696f28913fd59c02eefb975647c58f84ae718a66e65ec52d07e5904dc8"} Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.600794 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.764045 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef059f7c-88cc-442c-9f6e-5e7a5778a549-util\") pod \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\" (UID: \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\") " Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.764140 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef059f7c-88cc-442c-9f6e-5e7a5778a549-bundle\") pod \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\" (UID: \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\") " Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.764234 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8227\" (UniqueName: \"kubernetes.io/projected/ef059f7c-88cc-442c-9f6e-5e7a5778a549-kube-api-access-c8227\") pod \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\" (UID: \"ef059f7c-88cc-442c-9f6e-5e7a5778a549\") " Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.765182 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef059f7c-88cc-442c-9f6e-5e7a5778a549-bundle" (OuterVolumeSpecName: "bundle") pod "ef059f7c-88cc-442c-9f6e-5e7a5778a549" (UID: "ef059f7c-88cc-442c-9f6e-5e7a5778a549"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.770606 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef059f7c-88cc-442c-9f6e-5e7a5778a549-kube-api-access-c8227" (OuterVolumeSpecName: "kube-api-access-c8227") pod "ef059f7c-88cc-442c-9f6e-5e7a5778a549" (UID: "ef059f7c-88cc-442c-9f6e-5e7a5778a549"). InnerVolumeSpecName "kube-api-access-c8227". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.779160 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef059f7c-88cc-442c-9f6e-5e7a5778a549-util" (OuterVolumeSpecName: "util") pod "ef059f7c-88cc-442c-9f6e-5e7a5778a549" (UID: "ef059f7c-88cc-442c-9f6e-5e7a5778a549"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.866205 4931 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef059f7c-88cc-442c-9f6e-5e7a5778a549-util\") on node \"crc\" DevicePath \"\"" Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.866256 4931 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef059f7c-88cc-442c-9f6e-5e7a5778a549-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:15:01 crc kubenswrapper[4931]: I1129 00:15:01.866270 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8227\" (UniqueName: \"kubernetes.io/projected/ef059f7c-88cc-442c-9f6e-5e7a5778a549-kube-api-access-c8227\") on node \"crc\" DevicePath \"\"" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.097511 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.271335 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af0c698a-6e96-4837-80b1-de46e1c404f4-catalog-content\") pod \"af0c698a-6e96-4837-80b1-de46e1c404f4\" (UID: \"af0c698a-6e96-4837-80b1-de46e1c404f4\") " Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.271386 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpn9j\" (UniqueName: \"kubernetes.io/projected/af0c698a-6e96-4837-80b1-de46e1c404f4-kube-api-access-qpn9j\") pod \"af0c698a-6e96-4837-80b1-de46e1c404f4\" (UID: \"af0c698a-6e96-4837-80b1-de46e1c404f4\") " Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.271453 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af0c698a-6e96-4837-80b1-de46e1c404f4-utilities\") pod \"af0c698a-6e96-4837-80b1-de46e1c404f4\" (UID: \"af0c698a-6e96-4837-80b1-de46e1c404f4\") " Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.272367 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af0c698a-6e96-4837-80b1-de46e1c404f4-utilities" (OuterVolumeSpecName: "utilities") pod "af0c698a-6e96-4837-80b1-de46e1c404f4" (UID: "af0c698a-6e96-4837-80b1-de46e1c404f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.274920 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af0c698a-6e96-4837-80b1-de46e1c404f4-kube-api-access-qpn9j" (OuterVolumeSpecName: "kube-api-access-qpn9j") pod "af0c698a-6e96-4837-80b1-de46e1c404f4" (UID: "af0c698a-6e96-4837-80b1-de46e1c404f4"). InnerVolumeSpecName "kube-api-access-qpn9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.323658 4931 generic.go:334] "Generic (PLEG): container finished" podID="0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a" containerID="aa90dfd80e94e8ba9a47640a271f681b7a918f24890f3002c66d610ab88b01f2" exitCode=0 Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.323781 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" event={"ID":"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a","Type":"ContainerDied","Data":"aa90dfd80e94e8ba9a47640a271f681b7a918f24890f3002c66d610ab88b01f2"} Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.327950 4931 generic.go:334] "Generic (PLEG): container finished" podID="af0c698a-6e96-4837-80b1-de46e1c404f4" containerID="ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821" exitCode=0 Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.328016 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9xfht" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.328052 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfht" event={"ID":"af0c698a-6e96-4837-80b1-de46e1c404f4","Type":"ContainerDied","Data":"ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821"} Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.328102 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfht" event={"ID":"af0c698a-6e96-4837-80b1-de46e1c404f4","Type":"ContainerDied","Data":"b2b294e0bd9e2cd15511a11e7e053442144eeb604a60d292b0a5e6065310c1db"} Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.328122 4931 scope.go:117] "RemoveContainer" containerID="ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.332189 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" event={"ID":"ef059f7c-88cc-442c-9f6e-5e7a5778a549","Type":"ContainerDied","Data":"bc4864a139e95c085f8930011846d22bd38f8c756a3868f3de84d67638bd94ef"} Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.332259 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc4864a139e95c085f8930011846d22bd38f8c756a3868f3de84d67638bd94ef" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.332376 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.344066 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af0c698a-6e96-4837-80b1-de46e1c404f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af0c698a-6e96-4837-80b1-de46e1c404f4" (UID: "af0c698a-6e96-4837-80b1-de46e1c404f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.359290 4931 scope.go:117] "RemoveContainer" containerID="490077cd77ee808d0875c196c04a44e29bec3f6ba4d07251cca283f6f25180cf" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.373239 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af0c698a-6e96-4837-80b1-de46e1c404f4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.373319 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpn9j\" (UniqueName: \"kubernetes.io/projected/af0c698a-6e96-4837-80b1-de46e1c404f4-kube-api-access-qpn9j\") on node \"crc\" DevicePath \"\"" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.373341 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af0c698a-6e96-4837-80b1-de46e1c404f4-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.393535 4931 scope.go:117] "RemoveContainer" containerID="93a93893b6453f2db7f9fbf9e978bf652ba05d94b2b46ebadf978a146aa3e8fa" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.416883 4931 scope.go:117] "RemoveContainer" containerID="ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821" Nov 29 00:15:02 crc kubenswrapper[4931]: E1129 00:15:02.417837 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821\": container with ID starting with ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821 not found: ID does not exist" containerID="ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.417893 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821"} err="failed to get container status \"ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821\": rpc error: code = NotFound desc = could not find container \"ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821\": container with ID starting with ca8814dd0d02e19679c9bab5c3ba4f615c0c31be92a369f84d321640932fa821 not found: ID does not exist" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.417919 4931 scope.go:117] "RemoveContainer" containerID="490077cd77ee808d0875c196c04a44e29bec3f6ba4d07251cca283f6f25180cf" Nov 29 00:15:02 crc kubenswrapper[4931]: E1129 00:15:02.418564 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"490077cd77ee808d0875c196c04a44e29bec3f6ba4d07251cca283f6f25180cf\": container with ID starting with 490077cd77ee808d0875c196c04a44e29bec3f6ba4d07251cca283f6f25180cf not found: ID does not exist" containerID="490077cd77ee808d0875c196c04a44e29bec3f6ba4d07251cca283f6f25180cf" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.418622 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"490077cd77ee808d0875c196c04a44e29bec3f6ba4d07251cca283f6f25180cf"} err="failed to get container status \"490077cd77ee808d0875c196c04a44e29bec3f6ba4d07251cca283f6f25180cf\": rpc error: code = NotFound desc = could not find container \"490077cd77ee808d0875c196c04a44e29bec3f6ba4d07251cca283f6f25180cf\": container with ID starting with 490077cd77ee808d0875c196c04a44e29bec3f6ba4d07251cca283f6f25180cf not found: ID does not exist" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.418661 4931 scope.go:117] "RemoveContainer" containerID="93a93893b6453f2db7f9fbf9e978bf652ba05d94b2b46ebadf978a146aa3e8fa" Nov 29 00:15:02 crc kubenswrapper[4931]: E1129 00:15:02.419267 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93a93893b6453f2db7f9fbf9e978bf652ba05d94b2b46ebadf978a146aa3e8fa\": container with ID starting with 93a93893b6453f2db7f9fbf9e978bf652ba05d94b2b46ebadf978a146aa3e8fa not found: ID does not exist" containerID="93a93893b6453f2db7f9fbf9e978bf652ba05d94b2b46ebadf978a146aa3e8fa" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.419339 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93a93893b6453f2db7f9fbf9e978bf652ba05d94b2b46ebadf978a146aa3e8fa"} err="failed to get container status \"93a93893b6453f2db7f9fbf9e978bf652ba05d94b2b46ebadf978a146aa3e8fa\": rpc error: code = NotFound desc = could not find container \"93a93893b6453f2db7f9fbf9e978bf652ba05d94b2b46ebadf978a146aa3e8fa\": container with ID starting with 93a93893b6453f2db7f9fbf9e978bf652ba05d94b2b46ebadf978a146aa3e8fa not found: ID does not exist" Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.671568 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9xfht"] Nov 29 00:15:02 crc kubenswrapper[4931]: I1129 00:15:02.677914 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9xfht"] Nov 29 00:15:03 crc kubenswrapper[4931]: I1129 00:15:03.224239 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af0c698a-6e96-4837-80b1-de46e1c404f4" path="/var/lib/kubelet/pods/af0c698a-6e96-4837-80b1-de46e1c404f4/volumes" Nov 29 00:15:03 crc kubenswrapper[4931]: I1129 00:15:03.648634 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:03 crc kubenswrapper[4931]: I1129 00:15:03.695433 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmm6l\" (UniqueName: \"kubernetes.io/projected/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-kube-api-access-wmm6l\") pod \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\" (UID: \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\") " Nov 29 00:15:03 crc kubenswrapper[4931]: I1129 00:15:03.695526 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-secret-volume\") pod \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\" (UID: \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\") " Nov 29 00:15:03 crc kubenswrapper[4931]: I1129 00:15:03.695706 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-config-volume\") pod \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\" (UID: \"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a\") " Nov 29 00:15:03 crc kubenswrapper[4931]: I1129 00:15:03.697396 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-config-volume" (OuterVolumeSpecName: "config-volume") pod "0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a" (UID: "0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:15:03 crc kubenswrapper[4931]: I1129 00:15:03.702470 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a" (UID: "0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:15:03 crc kubenswrapper[4931]: I1129 00:15:03.703220 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-kube-api-access-wmm6l" (OuterVolumeSpecName: "kube-api-access-wmm6l") pod "0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a" (UID: "0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a"). InnerVolumeSpecName "kube-api-access-wmm6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:15:03 crc kubenswrapper[4931]: I1129 00:15:03.797228 4931 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 00:15:03 crc kubenswrapper[4931]: I1129 00:15:03.797280 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmm6l\" (UniqueName: \"kubernetes.io/projected/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-kube-api-access-wmm6l\") on node \"crc\" DevicePath \"\"" Nov 29 00:15:03 crc kubenswrapper[4931]: I1129 00:15:03.797303 4931 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 00:15:04 crc kubenswrapper[4931]: I1129 00:15:04.359443 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" event={"ID":"0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a","Type":"ContainerDied","Data":"596770696f28913fd59c02eefb975647c58f84ae718a66e65ec52d07e5904dc8"} Nov 29 00:15:04 crc kubenswrapper[4931]: I1129 00:15:04.359503 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="596770696f28913fd59c02eefb975647c58f84ae718a66e65ec52d07e5904dc8" Nov 29 00:15:04 crc kubenswrapper[4931]: I1129 00:15:04.359602 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.206538 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq"] Nov 29 00:15:07 crc kubenswrapper[4931]: E1129 00:15:07.207188 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a" containerName="collect-profiles" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.207207 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a" containerName="collect-profiles" Nov 29 00:15:07 crc kubenswrapper[4931]: E1129 00:15:07.207229 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef059f7c-88cc-442c-9f6e-5e7a5778a549" containerName="extract" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.207238 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef059f7c-88cc-442c-9f6e-5e7a5778a549" containerName="extract" Nov 29 00:15:07 crc kubenswrapper[4931]: E1129 00:15:07.207252 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0c698a-6e96-4837-80b1-de46e1c404f4" containerName="extract-content" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.207262 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0c698a-6e96-4837-80b1-de46e1c404f4" containerName="extract-content" Nov 29 00:15:07 crc kubenswrapper[4931]: E1129 00:15:07.207273 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0c698a-6e96-4837-80b1-de46e1c404f4" containerName="registry-server" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.207281 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0c698a-6e96-4837-80b1-de46e1c404f4" containerName="registry-server" Nov 29 00:15:07 crc kubenswrapper[4931]: E1129 00:15:07.207292 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef059f7c-88cc-442c-9f6e-5e7a5778a549" containerName="pull" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.207299 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef059f7c-88cc-442c-9f6e-5e7a5778a549" containerName="pull" Nov 29 00:15:07 crc kubenswrapper[4931]: E1129 00:15:07.207312 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0c698a-6e96-4837-80b1-de46e1c404f4" containerName="extract-utilities" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.207533 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0c698a-6e96-4837-80b1-de46e1c404f4" containerName="extract-utilities" Nov 29 00:15:07 crc kubenswrapper[4931]: E1129 00:15:07.207544 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef059f7c-88cc-442c-9f6e-5e7a5778a549" containerName="util" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.207553 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef059f7c-88cc-442c-9f6e-5e7a5778a549" containerName="util" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.207701 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef059f7c-88cc-442c-9f6e-5e7a5778a549" containerName="extract" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.207723 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="af0c698a-6e96-4837-80b1-de46e1c404f4" containerName="registry-server" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.207739 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a" containerName="collect-profiles" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.208254 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.210222 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-w5m47" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.222422 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq"] Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.244064 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tz4t\" (UniqueName: \"kubernetes.io/projected/fe758412-e16a-4c07-8c0d-e79e9f54cd73-kube-api-access-5tz4t\") pod \"openstack-operator-controller-operator-5757985b9c-rwxqq\" (UID: \"fe758412-e16a-4c07-8c0d-e79e9f54cd73\") " pod="openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.344858 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tz4t\" (UniqueName: \"kubernetes.io/projected/fe758412-e16a-4c07-8c0d-e79e9f54cd73-kube-api-access-5tz4t\") pod \"openstack-operator-controller-operator-5757985b9c-rwxqq\" (UID: \"fe758412-e16a-4c07-8c0d-e79e9f54cd73\") " pod="openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.361734 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tz4t\" (UniqueName: \"kubernetes.io/projected/fe758412-e16a-4c07-8c0d-e79e9f54cd73-kube-api-access-5tz4t\") pod \"openstack-operator-controller-operator-5757985b9c-rwxqq\" (UID: \"fe758412-e16a-4c07-8c0d-e79e9f54cd73\") " pod="openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.525170 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq" Nov 29 00:15:07 crc kubenswrapper[4931]: I1129 00:15:07.820184 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq"] Nov 29 00:15:08 crc kubenswrapper[4931]: I1129 00:15:08.397757 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq" event={"ID":"fe758412-e16a-4c07-8c0d-e79e9f54cd73","Type":"ContainerStarted","Data":"eb422be3f202782210e84e9a0ac4e9ade435250242a529f7434f6765eefb67bc"} Nov 29 00:15:13 crc kubenswrapper[4931]: I1129 00:15:13.437724 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq" event={"ID":"fe758412-e16a-4c07-8c0d-e79e9f54cd73","Type":"ContainerStarted","Data":"fc3d38cdb9fa0752a0137796814cb9aa4ed1a91f6da3d02a00ee9deaea65e775"} Nov 29 00:15:13 crc kubenswrapper[4931]: I1129 00:15:13.438597 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq" Nov 29 00:15:13 crc kubenswrapper[4931]: I1129 00:15:13.484452 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq" podStartSLOduration=1.9591779740000002 podStartE2EDuration="6.484433309s" podCreationTimestamp="2025-11-29 00:15:07 +0000 UTC" firstStartedPulling="2025-11-29 00:15:07.834051275 +0000 UTC m=+920.995944497" lastFinishedPulling="2025-11-29 00:15:12.3593066 +0000 UTC m=+925.521199832" observedRunningTime="2025-11-29 00:15:13.482073736 +0000 UTC m=+926.643967018" watchObservedRunningTime="2025-11-29 00:15:13.484433309 +0000 UTC m=+926.646326551" Nov 29 00:15:17 crc kubenswrapper[4931]: I1129 00:15:17.527609 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5757985b9c-rwxqq" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.399529 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.400994 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.406940 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-4h9h6" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.409948 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.411235 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.416004 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-pp65d" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.416431 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.417552 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.419723 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-n5cv7" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.424683 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.428601 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.442673 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.452753 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.454000 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.456613 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-6r58s" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.470238 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.474429 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.475608 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.479512 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-tpxzv" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.482519 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.499702 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.500562 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.508369 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-9wjwq" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.512409 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.521910 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.545006 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kjtn\" (UniqueName: \"kubernetes.io/projected/64f054e4-1ff5-4d20-a098-499b8baa25eb-kube-api-access-2kjtn\") pod \"barbican-operator-controller-manager-7d9dfd778-qkqqf\" (UID: \"64f054e4-1ff5-4d20-a098-499b8baa25eb\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.545084 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnwlz\" (UniqueName: \"kubernetes.io/projected/21842304-a783-4f75-bbb5-2ece9dacb6ca-kube-api-access-dnwlz\") pod \"heat-operator-controller-manager-5f64f6f8bb-nmtz7\" (UID: \"21842304-a783-4f75-bbb5-2ece9dacb6ca\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.545181 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkgfn\" (UniqueName: \"kubernetes.io/projected/b870bbdd-21a8-4279-9d2a-12dd26745325-kube-api-access-bkgfn\") pod \"glance-operator-controller-manager-668d9c48b9-lhc87\" (UID: \"b870bbdd-21a8-4279-9d2a-12dd26745325\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.545287 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj8sw\" (UniqueName: \"kubernetes.io/projected/256e71e1-5f65-4842-883e-1e3178e91965-kube-api-access-pj8sw\") pod \"designate-operator-controller-manager-78b4bc895b-wzvc6\" (UID: \"256e71e1-5f65-4842-883e-1e3178e91965\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.545359 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tqqs\" (UniqueName: \"kubernetes.io/projected/551c1c83-9dd7-47c7-8411-30881adc79bb-kube-api-access-6tqqs\") pod \"cinder-operator-controller-manager-859b6ccc6-tqw4f\" (UID: \"551c1c83-9dd7-47c7-8411-30881adc79bb\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.546572 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.554405 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.556892 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-fp9rr" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.559113 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.590824 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.590919 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.594070 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.600382 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-szlkv" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.604491 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.605466 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.614059 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-49fss" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.628380 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.629379 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.635423 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-l28ll" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.639365 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.653463 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.654150 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kjtn\" (UniqueName: \"kubernetes.io/projected/64f054e4-1ff5-4d20-a098-499b8baa25eb-kube-api-access-2kjtn\") pod \"barbican-operator-controller-manager-7d9dfd778-qkqqf\" (UID: \"64f054e4-1ff5-4d20-a098-499b8baa25eb\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.654185 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzrxs\" (UniqueName: \"kubernetes.io/projected/6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c-kube-api-access-pzrxs\") pod \"horizon-operator-controller-manager-68c6d99b8f-r5d5m\" (UID: \"6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.654222 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnwlz\" (UniqueName: \"kubernetes.io/projected/21842304-a783-4f75-bbb5-2ece9dacb6ca-kube-api-access-dnwlz\") pod \"heat-operator-controller-manager-5f64f6f8bb-nmtz7\" (UID: \"21842304-a783-4f75-bbb5-2ece9dacb6ca\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.654246 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lldtq\" (UniqueName: \"kubernetes.io/projected/3a157886-d4b5-4019-a0d6-0b40f6509b21-kube-api-access-lldtq\") pod \"infra-operator-controller-manager-57548d458d-nkmhr\" (UID: \"3a157886-d4b5-4019-a0d6-0b40f6509b21\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.654271 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkgfn\" (UniqueName: \"kubernetes.io/projected/b870bbdd-21a8-4279-9d2a-12dd26745325-kube-api-access-bkgfn\") pod \"glance-operator-controller-manager-668d9c48b9-lhc87\" (UID: \"b870bbdd-21a8-4279-9d2a-12dd26745325\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.654325 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj8sw\" (UniqueName: \"kubernetes.io/projected/256e71e1-5f65-4842-883e-1e3178e91965-kube-api-access-pj8sw\") pod \"designate-operator-controller-manager-78b4bc895b-wzvc6\" (UID: \"256e71e1-5f65-4842-883e-1e3178e91965\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.654352 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tqqs\" (UniqueName: \"kubernetes.io/projected/551c1c83-9dd7-47c7-8411-30881adc79bb-kube-api-access-6tqqs\") pod \"cinder-operator-controller-manager-859b6ccc6-tqw4f\" (UID: \"551c1c83-9dd7-47c7-8411-30881adc79bb\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.654389 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert\") pod \"infra-operator-controller-manager-57548d458d-nkmhr\" (UID: \"3a157886-d4b5-4019-a0d6-0b40f6509b21\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.689242 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.690829 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.692969 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tqqs\" (UniqueName: \"kubernetes.io/projected/551c1c83-9dd7-47c7-8411-30881adc79bb-kube-api-access-6tqqs\") pod \"cinder-operator-controller-manager-859b6ccc6-tqw4f\" (UID: \"551c1c83-9dd7-47c7-8411-30881adc79bb\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.693738 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnwlz\" (UniqueName: \"kubernetes.io/projected/21842304-a783-4f75-bbb5-2ece9dacb6ca-kube-api-access-dnwlz\") pod \"heat-operator-controller-manager-5f64f6f8bb-nmtz7\" (UID: \"21842304-a783-4f75-bbb5-2ece9dacb6ca\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.694066 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj8sw\" (UniqueName: \"kubernetes.io/projected/256e71e1-5f65-4842-883e-1e3178e91965-kube-api-access-pj8sw\") pod \"designate-operator-controller-manager-78b4bc895b-wzvc6\" (UID: \"256e71e1-5f65-4842-883e-1e3178e91965\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.697212 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.701528 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkgfn\" (UniqueName: \"kubernetes.io/projected/b870bbdd-21a8-4279-9d2a-12dd26745325-kube-api-access-bkgfn\") pod \"glance-operator-controller-manager-668d9c48b9-lhc87\" (UID: \"b870bbdd-21a8-4279-9d2a-12dd26745325\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.706453 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kjtn\" (UniqueName: \"kubernetes.io/projected/64f054e4-1ff5-4d20-a098-499b8baa25eb-kube-api-access-2kjtn\") pod \"barbican-operator-controller-manager-7d9dfd778-qkqqf\" (UID: \"64f054e4-1ff5-4d20-a098-499b8baa25eb\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.707620 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-vp5nj" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.713889 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.714901 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.719133 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-5zpmb" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.719705 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.735189 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.740931 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.741377 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.742277 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.743929 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-5vsr9" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.745363 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.746238 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.748173 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-zlp75" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.751508 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.752970 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.755417 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzrxs\" (UniqueName: \"kubernetes.io/projected/6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c-kube-api-access-pzrxs\") pod \"horizon-operator-controller-manager-68c6d99b8f-r5d5m\" (UID: \"6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.755627 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lldtq\" (UniqueName: \"kubernetes.io/projected/3a157886-d4b5-4019-a0d6-0b40f6509b21-kube-api-access-lldtq\") pod \"infra-operator-controller-manager-57548d458d-nkmhr\" (UID: \"3a157886-d4b5-4019-a0d6-0b40f6509b21\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.755663 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxszj\" (UniqueName: \"kubernetes.io/projected/5f790baf-d77a-41dd-840c-f4d9e13420e1-kube-api-access-dxszj\") pod \"ironic-operator-controller-manager-6c548fd776-zb94f\" (UID: \"5f790baf-d77a-41dd-840c-f4d9e13420e1\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.755738 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvf8s\" (UniqueName: \"kubernetes.io/projected/5a12e787-bf00-4f2b-8760-9b9a5cad0834-kube-api-access-nvf8s\") pod \"manila-operator-controller-manager-6546668bfd-b4skl\" (UID: \"5a12e787-bf00-4f2b-8760-9b9a5cad0834\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.755756 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert\") pod \"infra-operator-controller-manager-57548d458d-nkmhr\" (UID: \"3a157886-d4b5-4019-a0d6-0b40f6509b21\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:15:49 crc kubenswrapper[4931]: E1129 00:15:49.755881 4931 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.755974 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmffx\" (UniqueName: \"kubernetes.io/projected/0ce6e577-25c0-4aee-b14f-47e02ebfbef9-kube-api-access-nmffx\") pod \"keystone-operator-controller-manager-546d4bdf48-g82st\" (UID: \"0ce6e577-25c0-4aee-b14f-47e02ebfbef9\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st" Nov 29 00:15:49 crc kubenswrapper[4931]: E1129 00:15:49.757482 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert podName:3a157886-d4b5-4019-a0d6-0b40f6509b21 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:50.257465475 +0000 UTC m=+963.419358707 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert") pod "infra-operator-controller-manager-57548d458d-nkmhr" (UID: "3a157886-d4b5-4019-a0d6-0b40f6509b21") : secret "infra-operator-webhook-server-cert" not found Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.758171 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.759315 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.761087 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.761362 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-289dv" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.768173 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.775112 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.777769 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lldtq\" (UniqueName: \"kubernetes.io/projected/3a157886-d4b5-4019-a0d6-0b40f6509b21-kube-api-access-lldtq\") pod \"infra-operator-controller-manager-57548d458d-nkmhr\" (UID: \"3a157886-d4b5-4019-a0d6-0b40f6509b21\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.779766 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzrxs\" (UniqueName: \"kubernetes.io/projected/6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c-kube-api-access-pzrxs\") pod \"horizon-operator-controller-manager-68c6d99b8f-r5d5m\" (UID: \"6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.794105 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.796655 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-lc47s"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.799487 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-lc47s" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.801060 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-5kkjc" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.807794 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.817298 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.818456 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.818875 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.825039 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-bqnc6" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.829915 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-lc47s"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.847051 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.863032 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx7vb\" (UniqueName: \"kubernetes.io/projected/fd6606d5-8cd5-4308-a010-eea75cb666f1-kube-api-access-qx7vb\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp\" (UID: \"fd6606d5-8cd5-4308-a010-eea75cb666f1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.863320 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvf8s\" (UniqueName: \"kubernetes.io/projected/5a12e787-bf00-4f2b-8760-9b9a5cad0834-kube-api-access-nvf8s\") pod \"manila-operator-controller-manager-6546668bfd-b4skl\" (UID: \"5a12e787-bf00-4f2b-8760-9b9a5cad0834\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.863342 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jr8f\" (UniqueName: \"kubernetes.io/projected/724defda-f0a4-4edd-8a1a-7f6625893fe7-kube-api-access-2jr8f\") pod \"mariadb-operator-controller-manager-56bbcc9d85-2dbgl\" (UID: \"724defda-f0a4-4edd-8a1a-7f6625893fe7\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.863364 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9mk9\" (UniqueName: \"kubernetes.io/projected/365eb479-0132-4a0c-99d3-01cb923a9d61-kube-api-access-d9mk9\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-q7q7q\" (UID: \"365eb479-0132-4a0c-99d3-01cb923a9d61\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.863381 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmffx\" (UniqueName: \"kubernetes.io/projected/0ce6e577-25c0-4aee-b14f-47e02ebfbef9-kube-api-access-nmffx\") pod \"keystone-operator-controller-manager-546d4bdf48-g82st\" (UID: \"0ce6e577-25c0-4aee-b14f-47e02ebfbef9\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.863599 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.864802 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.863402 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv8s9\" (UniqueName: \"kubernetes.io/projected/c352f7a8-72dc-41da-9673-7fccd1a0974f-kube-api-access-xv8s9\") pod \"nova-operator-controller-manager-697bc559fc-lpf99\" (UID: \"c352f7a8-72dc-41da-9673-7fccd1a0974f\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.865246 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tfmc\" (UniqueName: \"kubernetes.io/projected/4c70c5cd-a329-4c90-b454-f836ddaef38f-kube-api-access-6tfmc\") pod \"octavia-operator-controller-manager-998648c74-z2mjc\" (UID: \"4c70c5cd-a329-4c90-b454-f836ddaef38f\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.865291 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxszj\" (UniqueName: \"kubernetes.io/projected/5f790baf-d77a-41dd-840c-f4d9e13420e1-kube-api-access-dxszj\") pod \"ironic-operator-controller-manager-6c548fd776-zb94f\" (UID: \"5f790baf-d77a-41dd-840c-f4d9e13420e1\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.865387 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp\" (UID: \"fd6606d5-8cd5-4308-a010-eea75cb666f1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.870602 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-wn5rz" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.888793 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmffx\" (UniqueName: \"kubernetes.io/projected/0ce6e577-25c0-4aee-b14f-47e02ebfbef9-kube-api-access-nmffx\") pod \"keystone-operator-controller-manager-546d4bdf48-g82st\" (UID: \"0ce6e577-25c0-4aee-b14f-47e02ebfbef9\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.891354 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvf8s\" (UniqueName: \"kubernetes.io/projected/5a12e787-bf00-4f2b-8760-9b9a5cad0834-kube-api-access-nvf8s\") pod \"manila-operator-controller-manager-6546668bfd-b4skl\" (UID: \"5a12e787-bf00-4f2b-8760-9b9a5cad0834\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.891421 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4"] Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.891942 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxszj\" (UniqueName: \"kubernetes.io/projected/5f790baf-d77a-41dd-840c-f4d9e13420e1-kube-api-access-dxszj\") pod \"ironic-operator-controller-manager-6c548fd776-zb94f\" (UID: \"5f790baf-d77a-41dd-840c-f4d9e13420e1\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.918449 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.923797 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.948411 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.967871 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp\" (UID: \"fd6606d5-8cd5-4308-a010-eea75cb666f1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.967937 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx7vb\" (UniqueName: \"kubernetes.io/projected/fd6606d5-8cd5-4308-a010-eea75cb666f1-kube-api-access-qx7vb\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp\" (UID: \"fd6606d5-8cd5-4308-a010-eea75cb666f1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.967991 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qznzk\" (UniqueName: \"kubernetes.io/projected/249844ae-a64c-4f9d-b204-bf3a1956aaa7-kube-api-access-qznzk\") pod \"swift-operator-controller-manager-5f8c65bbfc-f62b4\" (UID: \"249844ae-a64c-4f9d-b204-bf3a1956aaa7\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.968033 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jr8f\" (UniqueName: \"kubernetes.io/projected/724defda-f0a4-4edd-8a1a-7f6625893fe7-kube-api-access-2jr8f\") pod \"mariadb-operator-controller-manager-56bbcc9d85-2dbgl\" (UID: \"724defda-f0a4-4edd-8a1a-7f6625893fe7\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" Nov 29 00:15:49 crc kubenswrapper[4931]: E1129 00:15:49.968057 4931 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.968075 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9mk9\" (UniqueName: \"kubernetes.io/projected/365eb479-0132-4a0c-99d3-01cb923a9d61-kube-api-access-d9mk9\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-q7q7q\" (UID: \"365eb479-0132-4a0c-99d3-01cb923a9d61\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.968100 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv8s9\" (UniqueName: \"kubernetes.io/projected/c352f7a8-72dc-41da-9673-7fccd1a0974f-kube-api-access-xv8s9\") pod \"nova-operator-controller-manager-697bc559fc-lpf99\" (UID: \"c352f7a8-72dc-41da-9673-7fccd1a0974f\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99" Nov 29 00:15:49 crc kubenswrapper[4931]: E1129 00:15:49.968135 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert podName:fd6606d5-8cd5-4308-a010-eea75cb666f1 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:50.468112096 +0000 UTC m=+963.630005328 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" (UID: "fd6606d5-8cd5-4308-a010-eea75cb666f1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.968158 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tfmc\" (UniqueName: \"kubernetes.io/projected/4c70c5cd-a329-4c90-b454-f836ddaef38f-kube-api-access-6tfmc\") pod \"octavia-operator-controller-manager-998648c74-z2mjc\" (UID: \"4c70c5cd-a329-4c90-b454-f836ddaef38f\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.968203 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vbtq\" (UniqueName: \"kubernetes.io/projected/6f4cd8c8-caf0-4191-a3c4-51a172812b9b-kube-api-access-9vbtq\") pod \"ovn-operator-controller-manager-b6456fdb6-lnrrq\" (UID: \"6f4cd8c8-caf0-4191-a3c4-51a172812b9b\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" Nov 29 00:15:49 crc kubenswrapper[4931]: I1129 00:15:49.968250 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95rs4\" (UniqueName: \"kubernetes.io/projected/96c0b73e-c18e-4a52-a20c-e10db096b799-kube-api-access-95rs4\") pod \"placement-operator-controller-manager-78f8948974-lc47s\" (UID: \"96c0b73e-c18e-4a52-a20c-e10db096b799\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-lc47s" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.005614 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv8s9\" (UniqueName: \"kubernetes.io/projected/c352f7a8-72dc-41da-9673-7fccd1a0974f-kube-api-access-xv8s9\") pod \"nova-operator-controller-manager-697bc559fc-lpf99\" (UID: \"c352f7a8-72dc-41da-9673-7fccd1a0974f\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.006840 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jr8f\" (UniqueName: \"kubernetes.io/projected/724defda-f0a4-4edd-8a1a-7f6625893fe7-kube-api-access-2jr8f\") pod \"mariadb-operator-controller-manager-56bbcc9d85-2dbgl\" (UID: \"724defda-f0a4-4edd-8a1a-7f6625893fe7\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.009323 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9mk9\" (UniqueName: \"kubernetes.io/projected/365eb479-0132-4a0c-99d3-01cb923a9d61-kube-api-access-d9mk9\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-q7q7q\" (UID: \"365eb479-0132-4a0c-99d3-01cb923a9d61\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.017529 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tfmc\" (UniqueName: \"kubernetes.io/projected/4c70c5cd-a329-4c90-b454-f836ddaef38f-kube-api-access-6tfmc\") pod \"octavia-operator-controller-manager-998648c74-z2mjc\" (UID: \"4c70c5cd-a329-4c90-b454-f836ddaef38f\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.017701 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx7vb\" (UniqueName: \"kubernetes.io/projected/fd6606d5-8cd5-4308-a010-eea75cb666f1-kube-api-access-qx7vb\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp\" (UID: \"fd6606d5-8cd5-4308-a010-eea75cb666f1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.026627 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.029555 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.046160 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-tmkjs" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.073444 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vbtq\" (UniqueName: \"kubernetes.io/projected/6f4cd8c8-caf0-4191-a3c4-51a172812b9b-kube-api-access-9vbtq\") pod \"ovn-operator-controller-manager-b6456fdb6-lnrrq\" (UID: \"6f4cd8c8-caf0-4191-a3c4-51a172812b9b\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.073521 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95rs4\" (UniqueName: \"kubernetes.io/projected/96c0b73e-c18e-4a52-a20c-e10db096b799-kube-api-access-95rs4\") pod \"placement-operator-controller-manager-78f8948974-lc47s\" (UID: \"96c0b73e-c18e-4a52-a20c-e10db096b799\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-lc47s" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.073746 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qznzk\" (UniqueName: \"kubernetes.io/projected/249844ae-a64c-4f9d-b204-bf3a1956aaa7-kube-api-access-qznzk\") pod \"swift-operator-controller-manager-5f8c65bbfc-f62b4\" (UID: \"249844ae-a64c-4f9d-b204-bf3a1956aaa7\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.085502 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.098974 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.111263 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.116530 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.121481 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vbtq\" (UniqueName: \"kubernetes.io/projected/6f4cd8c8-caf0-4191-a3c4-51a172812b9b-kube-api-access-9vbtq\") pod \"ovn-operator-controller-manager-b6456fdb6-lnrrq\" (UID: \"6f4cd8c8-caf0-4191-a3c4-51a172812b9b\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.129125 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.129188 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95rs4\" (UniqueName: \"kubernetes.io/projected/96c0b73e-c18e-4a52-a20c-e10db096b799-kube-api-access-95rs4\") pod \"placement-operator-controller-manager-78f8948974-lc47s\" (UID: \"96c0b73e-c18e-4a52-a20c-e10db096b799\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-lc47s" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.129340 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qznzk\" (UniqueName: \"kubernetes.io/projected/249844ae-a64c-4f9d-b204-bf3a1956aaa7-kube-api-access-qznzk\") pod \"swift-operator-controller-manager-5f8c65bbfc-f62b4\" (UID: \"249844ae-a64c-4f9d-b204-bf3a1956aaa7\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.157023 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-dcl69"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.158346 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.161512 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-lc47s" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.163048 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-85kxv" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.169237 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.180620 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxks7\" (UniqueName: \"kubernetes.io/projected/b57413c5-3888-496c-ad0b-18128019b945-kube-api-access-sxks7\") pod \"telemetry-operator-controller-manager-76cc84c6bb-lthd8\" (UID: \"b57413c5-3888-496c-ad0b-18128019b945\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.216174 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.228033 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-dcl69"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.242657 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.243771 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.246714 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-sxq2l" Nov 29 00:15:50 crc kubenswrapper[4931]: W1129 00:15:50.251993 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64f054e4_1ff5_4d20_a098_499b8baa25eb.slice/crio-8bce39d8826b893e460901cac8b6cc76ff65639e4b747981f8bd919dc8b18cfe WatchSource:0}: Error finding container 8bce39d8826b893e460901cac8b6cc76ff65639e4b747981f8bd919dc8b18cfe: Status 404 returned error can't find the container with id 8bce39d8826b893e460901cac8b6cc76ff65639e4b747981f8bd919dc8b18cfe Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.265431 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.277040 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.279985 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.283874 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.284658 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert\") pod \"infra-operator-controller-manager-57548d458d-nkmhr\" (UID: \"3a157886-d4b5-4019-a0d6-0b40f6509b21\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.284703 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqbpv\" (UniqueName: \"kubernetes.io/projected/9621ec40-d523-423e-9bbe-4b063355cf08-kube-api-access-gqbpv\") pod \"test-operator-controller-manager-5854674fcc-dcl69\" (UID: \"9621ec40-d523-423e-9bbe-4b063355cf08\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.284740 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxks7\" (UniqueName: \"kubernetes.io/projected/b57413c5-3888-496c-ad0b-18128019b945-kube-api-access-sxks7\") pod \"telemetry-operator-controller-manager-76cc84c6bb-lthd8\" (UID: \"b57413c5-3888-496c-ad0b-18128019b945\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.285077 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.285113 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 29 00:15:50 crc kubenswrapper[4931]: E1129 00:15:50.285267 4931 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.285440 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-knn7q" Nov 29 00:15:50 crc kubenswrapper[4931]: E1129 00:15:50.285553 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert podName:3a157886-d4b5-4019-a0d6-0b40f6509b21 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:51.285536467 +0000 UTC m=+964.447429699 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert") pod "infra-operator-controller-manager-57548d458d-nkmhr" (UID: "3a157886-d4b5-4019-a0d6-0b40f6509b21") : secret "infra-operator-webhook-server-cert" not found Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.300035 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.301572 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.306523 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-9gkhx" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.306684 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.310554 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.353943 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxks7\" (UniqueName: \"kubernetes.io/projected/b57413c5-3888-496c-ad0b-18128019b945-kube-api-access-sxks7\") pod \"telemetry-operator-controller-manager-76cc84c6bb-lthd8\" (UID: \"b57413c5-3888-496c-ad0b-18128019b945\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.368188 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.400978 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.401279 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.401338 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxgg2\" (UniqueName: \"kubernetes.io/projected/0c61786d-9cf4-4447-82f7-47cc17ac65d8-kube-api-access-mxgg2\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.401358 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h8zn\" (UniqueName: \"kubernetes.io/projected/22500bd3-4958-4ba1-9b59-deb3032aea18-kube-api-access-5h8zn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-4qjb4\" (UID: \"22500bd3-4958-4ba1-9b59-deb3032aea18\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.401396 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngjtm\" (UniqueName: \"kubernetes.io/projected/2792db7c-cdd1-4356-97d5-fdba06bd44e4-kube-api-access-ngjtm\") pod \"watcher-operator-controller-manager-769dc69bc-qp9ht\" (UID: \"2792db7c-cdd1-4356-97d5-fdba06bd44e4\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.401415 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqbpv\" (UniqueName: \"kubernetes.io/projected/9621ec40-d523-423e-9bbe-4b063355cf08-kube-api-access-gqbpv\") pod \"test-operator-controller-manager-5854674fcc-dcl69\" (UID: \"9621ec40-d523-423e-9bbe-4b063355cf08\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.430026 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.436864 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqbpv\" (UniqueName: \"kubernetes.io/projected/9621ec40-d523-423e-9bbe-4b063355cf08-kube-api-access-gqbpv\") pod \"test-operator-controller-manager-5854674fcc-dcl69\" (UID: \"9621ec40-d523-423e-9bbe-4b063355cf08\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.490843 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.502069 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngjtm\" (UniqueName: \"kubernetes.io/projected/2792db7c-cdd1-4356-97d5-fdba06bd44e4-kube-api-access-ngjtm\") pod \"watcher-operator-controller-manager-769dc69bc-qp9ht\" (UID: \"2792db7c-cdd1-4356-97d5-fdba06bd44e4\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.502131 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.502149 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.502184 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp\" (UID: \"fd6606d5-8cd5-4308-a010-eea75cb666f1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.502205 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxgg2\" (UniqueName: \"kubernetes.io/projected/0c61786d-9cf4-4447-82f7-47cc17ac65d8-kube-api-access-mxgg2\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.502222 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h8zn\" (UniqueName: \"kubernetes.io/projected/22500bd3-4958-4ba1-9b59-deb3032aea18-kube-api-access-5h8zn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-4qjb4\" (UID: \"22500bd3-4958-4ba1-9b59-deb3032aea18\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4" Nov 29 00:15:50 crc kubenswrapper[4931]: E1129 00:15:50.502530 4931 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 00:15:50 crc kubenswrapper[4931]: E1129 00:15:50.502549 4931 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 29 00:15:50 crc kubenswrapper[4931]: E1129 00:15:50.502569 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert podName:fd6606d5-8cd5-4308-a010-eea75cb666f1 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:51.502555369 +0000 UTC m=+964.664448601 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" (UID: "fd6606d5-8cd5-4308-a010-eea75cb666f1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 00:15:50 crc kubenswrapper[4931]: E1129 00:15:50.502613 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs podName:0c61786d-9cf4-4447-82f7-47cc17ac65d8 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:51.0025954 +0000 UTC m=+964.164488632 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs") pod "openstack-operator-controller-manager-74fb9455b7-sz2k9" (UID: "0c61786d-9cf4-4447-82f7-47cc17ac65d8") : secret "metrics-server-cert" not found Nov 29 00:15:50 crc kubenswrapper[4931]: E1129 00:15:50.502874 4931 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 29 00:15:50 crc kubenswrapper[4931]: E1129 00:15:50.502899 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs podName:0c61786d-9cf4-4447-82f7-47cc17ac65d8 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:51.002890768 +0000 UTC m=+964.164784000 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs") pod "openstack-operator-controller-manager-74fb9455b7-sz2k9" (UID: "0c61786d-9cf4-4447-82f7-47cc17ac65d8") : secret "webhook-server-cert" not found Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.527185 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxgg2\" (UniqueName: \"kubernetes.io/projected/0c61786d-9cf4-4447-82f7-47cc17ac65d8-kube-api-access-mxgg2\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.527566 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngjtm\" (UniqueName: \"kubernetes.io/projected/2792db7c-cdd1-4356-97d5-fdba06bd44e4-kube-api-access-ngjtm\") pod \"watcher-operator-controller-manager-769dc69bc-qp9ht\" (UID: \"2792db7c-cdd1-4356-97d5-fdba06bd44e4\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.534212 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h8zn\" (UniqueName: \"kubernetes.io/projected/22500bd3-4958-4ba1-9b59-deb3032aea18-kube-api-access-5h8zn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-4qjb4\" (UID: \"22500bd3-4958-4ba1-9b59-deb3032aea18\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.600284 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.610084 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m"] Nov 29 00:15:50 crc kubenswrapper[4931]: W1129 00:15:50.625385 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c3cd3e3_0977_4a0d_8a36_35bfe26ab12c.slice/crio-8db393b23dc55c2527c6efe380721a13a5cd79e5e0ff7f2f2c8b371b377ddae9 WatchSource:0}: Error finding container 8db393b23dc55c2527c6efe380721a13a5cd79e5e0ff7f2f2c8b371b377ddae9: Status 404 returned error can't find the container with id 8db393b23dc55c2527c6efe380721a13a5cd79e5e0ff7f2f2c8b371b377ddae9 Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.671991 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4" Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.705530 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf" event={"ID":"64f054e4-1ff5-4d20-a098-499b8baa25eb","Type":"ContainerStarted","Data":"8bce39d8826b893e460901cac8b6cc76ff65639e4b747981f8bd919dc8b18cfe"} Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.706684 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6" event={"ID":"256e71e1-5f65-4842-883e-1e3178e91965","Type":"ContainerStarted","Data":"391a70622eec2683e480cd61c9637aaccef5c9c0d8124d2f5da216dc609db1a1"} Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.707986 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m" event={"ID":"6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c","Type":"ContainerStarted","Data":"8db393b23dc55c2527c6efe380721a13a5cd79e5e0ff7f2f2c8b371b377ddae9"} Nov 29 00:15:50 crc kubenswrapper[4931]: W1129 00:15:50.770283 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod551c1c83_9dd7_47c7_8411_30881adc79bb.slice/crio-35a8b63acb768311267e85b73d5c81519ed1a7993123fe1f07cda0e02ff50e8d WatchSource:0}: Error finding container 35a8b63acb768311267e85b73d5c81519ed1a7993123fe1f07cda0e02ff50e8d: Status 404 returned error can't find the container with id 35a8b63acb768311267e85b73d5c81519ed1a7993123fe1f07cda0e02ff50e8d Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.771551 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.777206 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.837276 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.865365 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.870385 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl"] Nov 29 00:15:50 crc kubenswrapper[4931]: I1129 00:15:50.884023 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f"] Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.001614 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl"] Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.009959 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.010092 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.010336 4931 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.010456 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs podName:0c61786d-9cf4-4447-82f7-47cc17ac65d8 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:52.01043749 +0000 UTC m=+965.172330882 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs") pod "openstack-operator-controller-manager-74fb9455b7-sz2k9" (UID: "0c61786d-9cf4-4447-82f7-47cc17ac65d8") : secret "webhook-server-cert" not found Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.012251 4931 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.012373 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs podName:0c61786d-9cf4-4447-82f7-47cc17ac65d8 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:52.012357941 +0000 UTC m=+965.174251173 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs") pod "openstack-operator-controller-manager-74fb9455b7-sz2k9" (UID: "0c61786d-9cf4-4447-82f7-47cc17ac65d8") : secret "metrics-server-cert" not found Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.019769 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99"] Nov 29 00:15:51 crc kubenswrapper[4931]: W1129 00:15:51.023695 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c70c5cd_a329_4c90_b454_f836ddaef38f.slice/crio-5ed4f2953d13ec17a9e34b9bb84fe31d7577ad96fa2a243bca8c0d4d1908703a WatchSource:0}: Error finding container 5ed4f2953d13ec17a9e34b9bb84fe31d7577ad96fa2a243bca8c0d4d1908703a: Status 404 returned error can't find the container with id 5ed4f2953d13ec17a9e34b9bb84fe31d7577ad96fa2a243bca8c0d4d1908703a Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.035069 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc"] Nov 29 00:15:51 crc kubenswrapper[4931]: W1129 00:15:51.039117 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96c0b73e_c18e_4a52_a20c_e10db096b799.slice/crio-f570d0eac0a82048f34fbc5dc5eeab1f7c792c116991bcc4f94a1be19898f681 WatchSource:0}: Error finding container f570d0eac0a82048f34fbc5dc5eeab1f7c792c116991bcc4f94a1be19898f681: Status 404 returned error can't find the container with id f570d0eac0a82048f34fbc5dc5eeab1f7c792c116991bcc4f94a1be19898f681 Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.041166 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-lc47s"] Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.083849 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8"] Nov 29 00:15:51 crc kubenswrapper[4931]: W1129 00:15:51.087559 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb57413c5_3888_496c_ad0b_18128019b945.slice/crio-ee88f31170c7ddb27f457e052c10a635d9f780d08745d1e331e888d5ea093e54 WatchSource:0}: Error finding container ee88f31170c7ddb27f457e052c10a635d9f780d08745d1e331e888d5ea093e54: Status 404 returned error can't find the container with id ee88f31170c7ddb27f457e052c10a635d9f780d08745d1e331e888d5ea093e54 Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.088832 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sxks7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-lthd8_openstack-operators(b57413c5-3888-496c-ad0b-18128019b945): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.090580 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sxks7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-lthd8_openstack-operators(b57413c5-3888-496c-ad0b-18128019b945): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.091916 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" podUID="b57413c5-3888-496c-ad0b-18128019b945" Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.109115 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q"] Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.126450 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq"] Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.133001 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9vbtq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-lnrrq_openstack-operators(6f4cd8c8-caf0-4191-a3c4-51a172812b9b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.134897 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9vbtq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-lnrrq_openstack-operators(6f4cd8c8-caf0-4191-a3c4-51a172812b9b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.136019 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" podUID="6f4cd8c8-caf0-4191-a3c4-51a172812b9b" Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.136435 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4"] Nov 29 00:15:51 crc kubenswrapper[4931]: W1129 00:15:51.138924 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod249844ae_a64c_4f9d_b204_bf3a1956aaa7.slice/crio-34d932163c68ff39f39ab28b61bc2c993d66d49ce913b1ec02635f342ae0fcf1 WatchSource:0}: Error finding container 34d932163c68ff39f39ab28b61bc2c993d66d49ce913b1ec02635f342ae0fcf1: Status 404 returned error can't find the container with id 34d932163c68ff39f39ab28b61bc2c993d66d49ce913b1ec02635f342ae0fcf1 Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.140749 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qznzk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-f62b4_openstack-operators(249844ae-a64c-4f9d-b204-bf3a1956aaa7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.143510 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qznzk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-f62b4_openstack-operators(249844ae-a64c-4f9d-b204-bf3a1956aaa7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.144676 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" podUID="249844ae-a64c-4f9d-b204-bf3a1956aaa7" Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.249326 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-dcl69"] Nov 29 00:15:51 crc kubenswrapper[4931]: W1129 00:15:51.249569 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9621ec40_d523_423e_9bbe_4b063355cf08.slice/crio-8fbf58e7046485d5e4137d84bb717f69f639f0e247c1fda421f12e380e28d574 WatchSource:0}: Error finding container 8fbf58e7046485d5e4137d84bb717f69f639f0e247c1fda421f12e380e28d574: Status 404 returned error can't find the container with id 8fbf58e7046485d5e4137d84bb717f69f639f0e247c1fda421f12e380e28d574 Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.251829 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqbpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-dcl69_openstack-operators(9621ec40-d523-423e-9bbe-4b063355cf08): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.253771 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqbpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-dcl69_openstack-operators(9621ec40-d523-423e-9bbe-4b063355cf08): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.254990 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" podUID="9621ec40-d523-423e-9bbe-4b063355cf08" Nov 29 00:15:51 crc kubenswrapper[4931]: W1129 00:15:51.255830 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2792db7c_cdd1_4356_97d5_fdba06bd44e4.slice/crio-efb3970eb4261d539089907546d13cda4c5fabdaacd9a52477795829356852e4 WatchSource:0}: Error finding container efb3970eb4261d539089907546d13cda4c5fabdaacd9a52477795829356852e4: Status 404 returned error can't find the container with id efb3970eb4261d539089907546d13cda4c5fabdaacd9a52477795829356852e4 Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.257435 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4"] Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.259066 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ngjtm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-qp9ht_openstack-operators(2792db7c-cdd1-4356-97d5-fdba06bd44e4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.260885 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ngjtm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-qp9ht_openstack-operators(2792db7c-cdd1-4356-97d5-fdba06bd44e4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.262063 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" podUID="2792db7c-cdd1-4356-97d5-fdba06bd44e4" Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.263651 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht"] Nov 29 00:15:51 crc kubenswrapper[4931]: W1129 00:15:51.264064 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22500bd3_4958_4ba1_9b59_deb3032aea18.slice/crio-796b4dc8ba1d9ba02d8da6cfc7aba8b175ec973c2fc2c5b426edc1f719352de5 WatchSource:0}: Error finding container 796b4dc8ba1d9ba02d8da6cfc7aba8b175ec973c2fc2c5b426edc1f719352de5: Status 404 returned error can't find the container with id 796b4dc8ba1d9ba02d8da6cfc7aba8b175ec973c2fc2c5b426edc1f719352de5 Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.268276 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5h8zn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-4qjb4_openstack-operators(22500bd3-4958-4ba1-9b59-deb3032aea18): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.270165 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4" podUID="22500bd3-4958-4ba1-9b59-deb3032aea18" Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.315671 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert\") pod \"infra-operator-controller-manager-57548d458d-nkmhr\" (UID: \"3a157886-d4b5-4019-a0d6-0b40f6509b21\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.316124 4931 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.316206 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert podName:3a157886-d4b5-4019-a0d6-0b40f6509b21 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:53.316187678 +0000 UTC m=+966.478080900 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert") pod "infra-operator-controller-manager-57548d458d-nkmhr" (UID: "3a157886-d4b5-4019-a0d6-0b40f6509b21") : secret "infra-operator-webhook-server-cert" not found Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.517668 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp\" (UID: \"fd6606d5-8cd5-4308-a010-eea75cb666f1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.517823 4931 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.517879 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert podName:fd6606d5-8cd5-4308-a010-eea75cb666f1 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:53.517865439 +0000 UTC m=+966.679758671 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" (UID: "fd6606d5-8cd5-4308-a010-eea75cb666f1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.716543 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4" event={"ID":"22500bd3-4958-4ba1-9b59-deb3032aea18","Type":"ContainerStarted","Data":"796b4dc8ba1d9ba02d8da6cfc7aba8b175ec973c2fc2c5b426edc1f719352de5"} Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.720598 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc" event={"ID":"4c70c5cd-a329-4c90-b454-f836ddaef38f","Type":"ContainerStarted","Data":"5ed4f2953d13ec17a9e34b9bb84fe31d7577ad96fa2a243bca8c0d4d1908703a"} Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.721176 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4" podUID="22500bd3-4958-4ba1-9b59-deb3032aea18" Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.722511 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" event={"ID":"b57413c5-3888-496c-ad0b-18128019b945","Type":"ContainerStarted","Data":"ee88f31170c7ddb27f457e052c10a635d9f780d08745d1e331e888d5ea093e54"} Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.724551 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" podUID="b57413c5-3888-496c-ad0b-18128019b945" Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.727590 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99" event={"ID":"c352f7a8-72dc-41da-9673-7fccd1a0974f","Type":"ContainerStarted","Data":"7c31c26a0a96ff3abd42ab69629d19bead4e4613fd3b9471b7bd9a9d7e0ca092"} Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.729107 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7" event={"ID":"21842304-a783-4f75-bbb5-2ece9dacb6ca","Type":"ContainerStarted","Data":"8b24dca05be7b94e880ed9b0d5b6215f617ea4fa3d0933319296c2003c8c93b3"} Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.734776 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" event={"ID":"6f4cd8c8-caf0-4191-a3c4-51a172812b9b","Type":"ContainerStarted","Data":"23aada3c62d1c7d328e364558c8e1e6529f588a6724f05c3b90ab5068d102f0b"} Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.737053 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" podUID="6f4cd8c8-caf0-4191-a3c4-51a172812b9b" Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.737785 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f" event={"ID":"551c1c83-9dd7-47c7-8411-30881adc79bb","Type":"ContainerStarted","Data":"35a8b63acb768311267e85b73d5c81519ed1a7993123fe1f07cda0e02ff50e8d"} Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.739331 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" event={"ID":"9621ec40-d523-423e-9bbe-4b063355cf08","Type":"ContainerStarted","Data":"8fbf58e7046485d5e4137d84bb717f69f639f0e247c1fda421f12e380e28d574"} Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.741475 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" event={"ID":"249844ae-a64c-4f9d-b204-bf3a1956aaa7","Type":"ContainerStarted","Data":"34d932163c68ff39f39ab28b61bc2c993d66d49ce913b1ec02635f342ae0fcf1"} Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.742218 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" podUID="9621ec40-d523-423e-9bbe-4b063355cf08" Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.746135 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" podUID="249844ae-a64c-4f9d-b204-bf3a1956aaa7" Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.763673 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st" event={"ID":"0ce6e577-25c0-4aee-b14f-47e02ebfbef9","Type":"ContainerStarted","Data":"c97fd801d96856b4b94ff4faaf29ca3db1adc4fd5932ab71cecf99b78617e4c9"} Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.764964 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" event={"ID":"365eb479-0132-4a0c-99d3-01cb923a9d61","Type":"ContainerStarted","Data":"7e70e76887bb4cf1396750828a385320723ff38924e950d48bd9d9ca57da0db0"} Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.774498 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f" event={"ID":"5f790baf-d77a-41dd-840c-f4d9e13420e1","Type":"ContainerStarted","Data":"ceab41583bd32b8cd85c53715d7da21d32c6988ae82e2f318cff8eddcfb13cf7"} Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.780332 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87" event={"ID":"b870bbdd-21a8-4279-9d2a-12dd26745325","Type":"ContainerStarted","Data":"21be169e93a864fc276d30a91985e62d6158c113c3561c29afd733ae58a47de9"} Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.791301 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-lc47s" event={"ID":"96c0b73e-c18e-4a52-a20c-e10db096b799","Type":"ContainerStarted","Data":"f570d0eac0a82048f34fbc5dc5eeab1f7c792c116991bcc4f94a1be19898f681"} Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.800037 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" event={"ID":"724defda-f0a4-4edd-8a1a-7f6625893fe7","Type":"ContainerStarted","Data":"490b5e97eb9710a45ec090351a95faec0fb66019a1c8ebadf2a39fad6d130d6b"} Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.801051 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" event={"ID":"5a12e787-bf00-4f2b-8760-9b9a5cad0834","Type":"ContainerStarted","Data":"e130395d0d0913bd8c6c6d24af9ed876df10cd25b4dbf0434591a95410049adc"} Nov 29 00:15:51 crc kubenswrapper[4931]: I1129 00:15:51.804247 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" event={"ID":"2792db7c-cdd1-4356-97d5-fdba06bd44e4","Type":"ContainerStarted","Data":"efb3970eb4261d539089907546d13cda4c5fabdaacd9a52477795829356852e4"} Nov 29 00:15:51 crc kubenswrapper[4931]: E1129 00:15:51.806627 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" podUID="2792db7c-cdd1-4356-97d5-fdba06bd44e4" Nov 29 00:15:52 crc kubenswrapper[4931]: I1129 00:15:52.024428 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:52 crc kubenswrapper[4931]: I1129 00:15:52.024566 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:52 crc kubenswrapper[4931]: E1129 00:15:52.024630 4931 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 29 00:15:52 crc kubenswrapper[4931]: E1129 00:15:52.024666 4931 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 29 00:15:52 crc kubenswrapper[4931]: E1129 00:15:52.024717 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs podName:0c61786d-9cf4-4447-82f7-47cc17ac65d8 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:54.024701892 +0000 UTC m=+967.186595124 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs") pod "openstack-operator-controller-manager-74fb9455b7-sz2k9" (UID: "0c61786d-9cf4-4447-82f7-47cc17ac65d8") : secret "webhook-server-cert" not found Nov 29 00:15:52 crc kubenswrapper[4931]: E1129 00:15:52.024730 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs podName:0c61786d-9cf4-4447-82f7-47cc17ac65d8 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:54.024724622 +0000 UTC m=+967.186617854 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs") pod "openstack-operator-controller-manager-74fb9455b7-sz2k9" (UID: "0c61786d-9cf4-4447-82f7-47cc17ac65d8") : secret "metrics-server-cert" not found Nov 29 00:15:52 crc kubenswrapper[4931]: E1129 00:15:52.817551 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4" podUID="22500bd3-4958-4ba1-9b59-deb3032aea18" Nov 29 00:15:52 crc kubenswrapper[4931]: E1129 00:15:52.817601 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" podUID="249844ae-a64c-4f9d-b204-bf3a1956aaa7" Nov 29 00:15:52 crc kubenswrapper[4931]: E1129 00:15:52.818124 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" podUID="6f4cd8c8-caf0-4191-a3c4-51a172812b9b" Nov 29 00:15:52 crc kubenswrapper[4931]: E1129 00:15:52.819181 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" podUID="9621ec40-d523-423e-9bbe-4b063355cf08" Nov 29 00:15:52 crc kubenswrapper[4931]: E1129 00:15:52.819498 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" podUID="2792db7c-cdd1-4356-97d5-fdba06bd44e4" Nov 29 00:15:52 crc kubenswrapper[4931]: E1129 00:15:52.819942 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" podUID="b57413c5-3888-496c-ad0b-18128019b945" Nov 29 00:15:53 crc kubenswrapper[4931]: I1129 00:15:53.343277 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert\") pod \"infra-operator-controller-manager-57548d458d-nkmhr\" (UID: \"3a157886-d4b5-4019-a0d6-0b40f6509b21\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:15:53 crc kubenswrapper[4931]: E1129 00:15:53.343689 4931 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 29 00:15:53 crc kubenswrapper[4931]: E1129 00:15:53.343851 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert podName:3a157886-d4b5-4019-a0d6-0b40f6509b21 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:57.343821567 +0000 UTC m=+970.505714799 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert") pod "infra-operator-controller-manager-57548d458d-nkmhr" (UID: "3a157886-d4b5-4019-a0d6-0b40f6509b21") : secret "infra-operator-webhook-server-cert" not found Nov 29 00:15:53 crc kubenswrapper[4931]: I1129 00:15:53.545960 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp\" (UID: \"fd6606d5-8cd5-4308-a010-eea75cb666f1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:15:53 crc kubenswrapper[4931]: E1129 00:15:53.546180 4931 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 00:15:53 crc kubenswrapper[4931]: E1129 00:15:53.546266 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert podName:fd6606d5-8cd5-4308-a010-eea75cb666f1 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:57.546246857 +0000 UTC m=+970.708140089 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" (UID: "fd6606d5-8cd5-4308-a010-eea75cb666f1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 00:15:54 crc kubenswrapper[4931]: I1129 00:15:54.053511 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:54 crc kubenswrapper[4931]: I1129 00:15:54.054252 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:54 crc kubenswrapper[4931]: E1129 00:15:54.053652 4931 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 29 00:15:54 crc kubenswrapper[4931]: E1129 00:15:54.054489 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs podName:0c61786d-9cf4-4447-82f7-47cc17ac65d8 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:58.054468268 +0000 UTC m=+971.216361510 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs") pod "openstack-operator-controller-manager-74fb9455b7-sz2k9" (UID: "0c61786d-9cf4-4447-82f7-47cc17ac65d8") : secret "webhook-server-cert" not found Nov 29 00:15:54 crc kubenswrapper[4931]: E1129 00:15:54.054344 4931 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 29 00:15:54 crc kubenswrapper[4931]: E1129 00:15:54.056591 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs podName:0c61786d-9cf4-4447-82f7-47cc17ac65d8 nodeName:}" failed. No retries permitted until 2025-11-29 00:15:58.056493222 +0000 UTC m=+971.218386484 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs") pod "openstack-operator-controller-manager-74fb9455b7-sz2k9" (UID: "0c61786d-9cf4-4447-82f7-47cc17ac65d8") : secret "metrics-server-cert" not found Nov 29 00:15:57 crc kubenswrapper[4931]: I1129 00:15:57.404408 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert\") pod \"infra-operator-controller-manager-57548d458d-nkmhr\" (UID: \"3a157886-d4b5-4019-a0d6-0b40f6509b21\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:15:57 crc kubenswrapper[4931]: E1129 00:15:57.404603 4931 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 29 00:15:57 crc kubenswrapper[4931]: E1129 00:15:57.405201 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert podName:3a157886-d4b5-4019-a0d6-0b40f6509b21 nodeName:}" failed. No retries permitted until 2025-11-29 00:16:05.405180388 +0000 UTC m=+978.567073620 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert") pod "infra-operator-controller-manager-57548d458d-nkmhr" (UID: "3a157886-d4b5-4019-a0d6-0b40f6509b21") : secret "infra-operator-webhook-server-cert" not found Nov 29 00:15:57 crc kubenswrapper[4931]: I1129 00:15:57.611199 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp\" (UID: \"fd6606d5-8cd5-4308-a010-eea75cb666f1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:15:57 crc kubenswrapper[4931]: E1129 00:15:57.611397 4931 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 00:15:57 crc kubenswrapper[4931]: E1129 00:15:57.611454 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert podName:fd6606d5-8cd5-4308-a010-eea75cb666f1 nodeName:}" failed. No retries permitted until 2025-11-29 00:16:05.611437012 +0000 UTC m=+978.773330264 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" (UID: "fd6606d5-8cd5-4308-a010-eea75cb666f1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 29 00:15:58 crc kubenswrapper[4931]: I1129 00:15:58.117746 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:58 crc kubenswrapper[4931]: I1129 00:15:58.117824 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:15:58 crc kubenswrapper[4931]: E1129 00:15:58.117962 4931 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 29 00:15:58 crc kubenswrapper[4931]: E1129 00:15:58.118028 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs podName:0c61786d-9cf4-4447-82f7-47cc17ac65d8 nodeName:}" failed. No retries permitted until 2025-11-29 00:16:06.118010638 +0000 UTC m=+979.279903890 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs") pod "openstack-operator-controller-manager-74fb9455b7-sz2k9" (UID: "0c61786d-9cf4-4447-82f7-47cc17ac65d8") : secret "metrics-server-cert" not found Nov 29 00:15:58 crc kubenswrapper[4931]: E1129 00:15:58.118064 4931 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 29 00:15:58 crc kubenswrapper[4931]: E1129 00:15:58.118197 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs podName:0c61786d-9cf4-4447-82f7-47cc17ac65d8 nodeName:}" failed. No retries permitted until 2025-11-29 00:16:06.118166992 +0000 UTC m=+979.280060224 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs") pod "openstack-operator-controller-manager-74fb9455b7-sz2k9" (UID: "0c61786d-9cf4-4447-82f7-47cc17ac65d8") : secret "webhook-server-cert" not found Nov 29 00:16:02 crc kubenswrapper[4931]: E1129 00:16:02.538766 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nvf8s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6546668bfd-b4skl_openstack-operators(5a12e787-bf00-4f2b-8760-9b9a5cad0834): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:16:02 crc kubenswrapper[4931]: E1129 00:16:02.540585 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" podUID="5a12e787-bf00-4f2b-8760-9b9a5cad0834" Nov 29 00:16:02 crc kubenswrapper[4931]: E1129 00:16:02.545868 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d9mk9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-q7q7q_openstack-operators(365eb479-0132-4a0c-99d3-01cb923a9d61): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:16:02 crc kubenswrapper[4931]: E1129 00:16:02.547466 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" podUID="365eb479-0132-4a0c-99d3-01cb923a9d61" Nov 29 00:16:02 crc kubenswrapper[4931]: E1129 00:16:02.568769 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2jr8f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-2dbgl_openstack-operators(724defda-f0a4-4edd-8a1a-7f6625893fe7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 29 00:16:02 crc kubenswrapper[4931]: E1129 00:16:02.571679 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" podUID="724defda-f0a4-4edd-8a1a-7f6625893fe7" Nov 29 00:16:02 crc kubenswrapper[4931]: I1129 00:16:02.903673 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f" event={"ID":"551c1c83-9dd7-47c7-8411-30881adc79bb","Type":"ContainerStarted","Data":"27c5a8699fdaadb6107e4a1f0032862427d3cb5183cdc789d09cf4e5524d4df4"} Nov 29 00:16:02 crc kubenswrapper[4931]: I1129 00:16:02.924887 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc" event={"ID":"4c70c5cd-a329-4c90-b454-f836ddaef38f","Type":"ContainerStarted","Data":"45d197d05cbfac3010fce230dde236b18df8752ac1620831f54395f40e68519d"} Nov 29 00:16:02 crc kubenswrapper[4931]: I1129 00:16:02.953533 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" event={"ID":"365eb479-0132-4a0c-99d3-01cb923a9d61","Type":"ContainerStarted","Data":"501f4c6c89654c0ce612c613549918ab5938bd60d4eec7827dcea86aedf3a055"} Nov 29 00:16:02 crc kubenswrapper[4931]: I1129 00:16:02.954311 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" Nov 29 00:16:02 crc kubenswrapper[4931]: E1129 00:16:02.957501 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" podUID="365eb479-0132-4a0c-99d3-01cb923a9d61" Nov 29 00:16:02 crc kubenswrapper[4931]: I1129 00:16:02.960366 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7" event={"ID":"21842304-a783-4f75-bbb5-2ece9dacb6ca","Type":"ContainerStarted","Data":"4e1a313f6c82e202d3db23d4b1650f1013e71a32182d59eed90ec1c2af8a08c1"} Nov 29 00:16:02 crc kubenswrapper[4931]: I1129 00:16:02.962009 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf" event={"ID":"64f054e4-1ff5-4d20-a098-499b8baa25eb","Type":"ContainerStarted","Data":"96c35864cad4386de94bb5dd320964aeb3594a161110567bd22f06ee2a3f1f33"} Nov 29 00:16:02 crc kubenswrapper[4931]: I1129 00:16:02.973309 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6" event={"ID":"256e71e1-5f65-4842-883e-1e3178e91965","Type":"ContainerStarted","Data":"ce7a5246bebc96e0f2fe62a4050f6ca7ea140c0841c66dcc2443963c5334d7ee"} Nov 29 00:16:02 crc kubenswrapper[4931]: I1129 00:16:02.978988 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st" event={"ID":"0ce6e577-25c0-4aee-b14f-47e02ebfbef9","Type":"ContainerStarted","Data":"547f4c37cf0167b472a07859e6fa5c6cc496ce88328b56480dc3e497310a5bc9"} Nov 29 00:16:02 crc kubenswrapper[4931]: I1129 00:16:02.996604 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" event={"ID":"724defda-f0a4-4edd-8a1a-7f6625893fe7","Type":"ContainerStarted","Data":"d45c15d08d3fba4dcc67379c5aee072607114464487a03ce47371ea9496d73a7"} Nov 29 00:16:02 crc kubenswrapper[4931]: I1129 00:16:02.996925 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" Nov 29 00:16:03 crc kubenswrapper[4931]: E1129 00:16:02.999314 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" podUID="724defda-f0a4-4edd-8a1a-7f6625893fe7" Nov 29 00:16:03 crc kubenswrapper[4931]: I1129 00:16:03.006662 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" event={"ID":"5a12e787-bf00-4f2b-8760-9b9a5cad0834","Type":"ContainerStarted","Data":"a97eed24d6de424f0434c53a68ad3bc70f039bdda95307c118133395584ee469"} Nov 29 00:16:03 crc kubenswrapper[4931]: I1129 00:16:03.006965 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" Nov 29 00:16:03 crc kubenswrapper[4931]: E1129 00:16:03.041630 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" podUID="5a12e787-bf00-4f2b-8760-9b9a5cad0834" Nov 29 00:16:03 crc kubenswrapper[4931]: I1129 00:16:03.057638 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87" event={"ID":"b870bbdd-21a8-4279-9d2a-12dd26745325","Type":"ContainerStarted","Data":"e8accd2e66887ed2aaafe46fbcebb6dbd85c3f7ab0c07d7d56dd73b41ff7cda1"} Nov 29 00:16:03 crc kubenswrapper[4931]: I1129 00:16:03.060601 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m" event={"ID":"6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c","Type":"ContainerStarted","Data":"ac4ca7b1a36a2128999fe036a07304977b38c36d1768cc5e3239507900eff4d2"} Nov 29 00:16:03 crc kubenswrapper[4931]: I1129 00:16:03.062187 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f" event={"ID":"5f790baf-d77a-41dd-840c-f4d9e13420e1","Type":"ContainerStarted","Data":"798295570e12cba50e95a1bea24a62ca8ae551e4a3eae4579aa89246a3fd26dd"} Nov 29 00:16:03 crc kubenswrapper[4931]: I1129 00:16:03.086514 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-lc47s" event={"ID":"96c0b73e-c18e-4a52-a20c-e10db096b799","Type":"ContainerStarted","Data":"35c6f06dfae7a8a32e5e2cc7a6c19b53abf518b9aa3eb5945079771a607a0c25"} Nov 29 00:16:03 crc kubenswrapper[4931]: I1129 00:16:03.087939 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99" event={"ID":"c352f7a8-72dc-41da-9673-7fccd1a0974f","Type":"ContainerStarted","Data":"43cc8655702c55775936da50dd50df0d4cfbf00ec6eb7c6ff27990e043ea148e"} Nov 29 00:16:04 crc kubenswrapper[4931]: E1129 00:16:04.097460 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" podUID="365eb479-0132-4a0c-99d3-01cb923a9d61" Nov 29 00:16:04 crc kubenswrapper[4931]: E1129 00:16:04.097865 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" podUID="5a12e787-bf00-4f2b-8760-9b9a5cad0834" Nov 29 00:16:04 crc kubenswrapper[4931]: E1129 00:16:04.099090 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" podUID="724defda-f0a4-4edd-8a1a-7f6625893fe7" Nov 29 00:16:05 crc kubenswrapper[4931]: I1129 00:16:05.430554 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert\") pod \"infra-operator-controller-manager-57548d458d-nkmhr\" (UID: \"3a157886-d4b5-4019-a0d6-0b40f6509b21\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:16:05 crc kubenswrapper[4931]: I1129 00:16:05.452724 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a157886-d4b5-4019-a0d6-0b40f6509b21-cert\") pod \"infra-operator-controller-manager-57548d458d-nkmhr\" (UID: \"3a157886-d4b5-4019-a0d6-0b40f6509b21\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:16:05 crc kubenswrapper[4931]: I1129 00:16:05.501958 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:16:05 crc kubenswrapper[4931]: I1129 00:16:05.633475 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp\" (UID: \"fd6606d5-8cd5-4308-a010-eea75cb666f1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:16:05 crc kubenswrapper[4931]: I1129 00:16:05.637573 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd6606d5-8cd5-4308-a010-eea75cb666f1-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp\" (UID: \"fd6606d5-8cd5-4308-a010-eea75cb666f1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:16:05 crc kubenswrapper[4931]: I1129 00:16:05.739254 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:16:06 crc kubenswrapper[4931]: I1129 00:16:06.142054 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:16:06 crc kubenswrapper[4931]: I1129 00:16:06.142113 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:16:06 crc kubenswrapper[4931]: I1129 00:16:06.149318 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-metrics-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:16:06 crc kubenswrapper[4931]: I1129 00:16:06.149940 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0c61786d-9cf4-4447-82f7-47cc17ac65d8-webhook-certs\") pod \"openstack-operator-controller-manager-74fb9455b7-sz2k9\" (UID: \"0c61786d-9cf4-4447-82f7-47cc17ac65d8\") " pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:16:06 crc kubenswrapper[4931]: I1129 00:16:06.241467 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:16:06 crc kubenswrapper[4931]: I1129 00:16:06.626060 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp"] Nov 29 00:16:06 crc kubenswrapper[4931]: I1129 00:16:06.700664 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr"] Nov 29 00:16:06 crc kubenswrapper[4931]: W1129 00:16:06.835330 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a157886_d4b5_4019_a0d6_0b40f6509b21.slice/crio-121b511e307adb2819691d2931352d1a711eef75767c4b4288fc96d0880b5484 WatchSource:0}: Error finding container 121b511e307adb2819691d2931352d1a711eef75767c4b4288fc96d0880b5484: Status 404 returned error can't find the container with id 121b511e307adb2819691d2931352d1a711eef75767c4b4288fc96d0880b5484 Nov 29 00:16:07 crc kubenswrapper[4931]: I1129 00:16:07.133922 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" event={"ID":"3a157886-d4b5-4019-a0d6-0b40f6509b21","Type":"ContainerStarted","Data":"121b511e307adb2819691d2931352d1a711eef75767c4b4288fc96d0880b5484"} Nov 29 00:16:07 crc kubenswrapper[4931]: W1129 00:16:07.403910 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd6606d5_8cd5_4308_a010_eea75cb666f1.slice/crio-f46e95ab7c320432939c797ebd709f6a288c06e653452791770f1d1d830685e4 WatchSource:0}: Error finding container f46e95ab7c320432939c797ebd709f6a288c06e653452791770f1d1d830685e4: Status 404 returned error can't find the container with id f46e95ab7c320432939c797ebd709f6a288c06e653452791770f1d1d830685e4 Nov 29 00:16:07 crc kubenswrapper[4931]: I1129 00:16:07.615089 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9"] Nov 29 00:16:08 crc kubenswrapper[4931]: I1129 00:16:08.142289 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" event={"ID":"fd6606d5-8cd5-4308-a010-eea75cb666f1","Type":"ContainerStarted","Data":"f46e95ab7c320432939c797ebd709f6a288c06e653452791770f1d1d830685e4"} Nov 29 00:16:08 crc kubenswrapper[4931]: W1129 00:16:08.646159 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c61786d_9cf4_4447_82f7_47cc17ac65d8.slice/crio-ee240619fb782038b36b2650a5fcf6f790118f5b44eac503901813e4abd255d0 WatchSource:0}: Error finding container ee240619fb782038b36b2650a5fcf6f790118f5b44eac503901813e4abd255d0: Status 404 returned error can't find the container with id ee240619fb782038b36b2650a5fcf6f790118f5b44eac503901813e4abd255d0 Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.152876 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87" event={"ID":"b870bbdd-21a8-4279-9d2a-12dd26745325","Type":"ContainerStarted","Data":"e3c64b9e464b8558a508a0679044809715d61354ace8a2b8bacd97083aa0b572"} Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.153419 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.154839 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.155888 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf" event={"ID":"64f054e4-1ff5-4d20-a098-499b8baa25eb","Type":"ContainerStarted","Data":"7d93cf1d1602c3a8c3e62cc9b3eacdab2f7f5aa906b4675a5a8bbef743c9b7a5"} Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.156156 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.159715 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.160600 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6" event={"ID":"256e71e1-5f65-4842-883e-1e3178e91965","Type":"ContainerStarted","Data":"acf53dbc8095375757ca3efe91b1e15eb5cdeec3461875aee8fa192db68bc21b"} Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.160917 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.163293 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc" event={"ID":"4c70c5cd-a329-4c90-b454-f836ddaef38f","Type":"ContainerStarted","Data":"b18ddd0e6c75ae44297785443378c9c8a810d0bb1f22978715b5974f8aa93d11"} Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.163979 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.164873 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.165094 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.166734 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m" event={"ID":"6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c","Type":"ContainerStarted","Data":"2e12d28d01f6d0a156b608719c368243aee8fd93244c6a734057ead42f1b949a"} Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.166906 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.168203 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.168944 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f" event={"ID":"5f790baf-d77a-41dd-840c-f4d9e13420e1","Type":"ContainerStarted","Data":"21ddeb64164ec0c0239f5576fac27a9d237a82061086bd78e735581aae54b28d"} Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.169152 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.175754 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.178940 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-lhc87" podStartSLOduration=3.555730383 podStartE2EDuration="20.178924336s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:50.895889992 +0000 UTC m=+964.057783224" lastFinishedPulling="2025-11-29 00:16:07.519083935 +0000 UTC m=+980.680977177" observedRunningTime="2025-11-29 00:16:09.175409934 +0000 UTC m=+982.337303176" watchObservedRunningTime="2025-11-29 00:16:09.178924336 +0000 UTC m=+982.340817568" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.190625 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" event={"ID":"6f4cd8c8-caf0-4191-a3c4-51a172812b9b","Type":"ContainerStarted","Data":"ae8db4d859390567dbd74895b5a5904cc041d927929f1fd098b4e103c813859d"} Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.196073 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-z2mjc" podStartSLOduration=5.017617511 podStartE2EDuration="20.196059769s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:51.02540052 +0000 UTC m=+964.187293752" lastFinishedPulling="2025-11-29 00:16:06.203842788 +0000 UTC m=+979.365736010" observedRunningTime="2025-11-29 00:16:09.188575234 +0000 UTC m=+982.350468466" watchObservedRunningTime="2025-11-29 00:16:09.196059769 +0000 UTC m=+982.357953001" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.200600 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" event={"ID":"0c61786d-9cf4-4447-82f7-47cc17ac65d8","Type":"ContainerStarted","Data":"c1dcf3aada39f2f424417d2f735bc81d1968aba6f175a2e63642cec2f4893a72"} Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.200647 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" event={"ID":"0c61786d-9cf4-4447-82f7-47cc17ac65d8","Type":"ContainerStarted","Data":"ee240619fb782038b36b2650a5fcf6f790118f5b44eac503901813e4abd255d0"} Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.201237 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.206689 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" event={"ID":"249844ae-a64c-4f9d-b204-bf3a1956aaa7","Type":"ContainerStarted","Data":"b01470419a67f47cc37a91a7dc09e9da4a527f3f5d2baf45210713538d6b400b"} Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.207559 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wzvc6" podStartSLOduration=3.235551735 podStartE2EDuration="20.20754314s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:50.492691164 +0000 UTC m=+963.654584396" lastFinishedPulling="2025-11-29 00:16:07.464682569 +0000 UTC m=+980.626575801" observedRunningTime="2025-11-29 00:16:09.205553132 +0000 UTC m=+982.367446374" watchObservedRunningTime="2025-11-29 00:16:09.20754314 +0000 UTC m=+982.369436372" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.231524 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zb94f" podStartSLOduration=2.482911714 podStartE2EDuration="20.23150719s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:50.900918957 +0000 UTC m=+964.062812199" lastFinishedPulling="2025-11-29 00:16:08.649514413 +0000 UTC m=+981.811407675" observedRunningTime="2025-11-29 00:16:09.222446629 +0000 UTC m=+982.384339871" watchObservedRunningTime="2025-11-29 00:16:09.23150719 +0000 UTC m=+982.393400422" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.251590 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-qkqqf" podStartSLOduration=4.369503976 podStartE2EDuration="20.251574367s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:50.312050497 +0000 UTC m=+963.473943729" lastFinishedPulling="2025-11-29 00:16:06.194120888 +0000 UTC m=+979.356014120" observedRunningTime="2025-11-29 00:16:09.250208718 +0000 UTC m=+982.412101950" watchObservedRunningTime="2025-11-29 00:16:09.251574367 +0000 UTC m=+982.413467599" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.266916 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-r5d5m" podStartSLOduration=4.701115967 podStartE2EDuration="20.266891838s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:50.628343477 +0000 UTC m=+963.790236709" lastFinishedPulling="2025-11-29 00:16:06.194119348 +0000 UTC m=+979.356012580" observedRunningTime="2025-11-29 00:16:09.266373944 +0000 UTC m=+982.428267186" watchObservedRunningTime="2025-11-29 00:16:09.266891838 +0000 UTC m=+982.428785070" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.330684 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" podStartSLOduration=19.330668575 podStartE2EDuration="19.330668575s" podCreationTimestamp="2025-11-29 00:15:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:16:09.32565487 +0000 UTC m=+982.487548112" watchObservedRunningTime="2025-11-29 00:16:09.330668575 +0000 UTC m=+982.492561807" Nov 29 00:16:09 crc kubenswrapper[4931]: I1129 00:16:09.952205 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" Nov 29 00:16:10 crc kubenswrapper[4931]: I1129 00:16:10.103261 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" Nov 29 00:16:10 crc kubenswrapper[4931]: I1129 00:16:10.113699 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" Nov 29 00:16:11 crc kubenswrapper[4931]: I1129 00:16:11.224082 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" event={"ID":"9621ec40-d523-423e-9bbe-4b063355cf08","Type":"ContainerStarted","Data":"3da4473f350310226920209fe3573e3e31374a9e479c34142b2d3b3ea13d4495"} Nov 29 00:16:11 crc kubenswrapper[4931]: I1129 00:16:11.226246 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-lc47s" event={"ID":"96c0b73e-c18e-4a52-a20c-e10db096b799","Type":"ContainerStarted","Data":"b7334b31f8d9edd47d8778fc44a7d654b8ec4bcbaa7ea7b8a53e9f001d7f7119"} Nov 29 00:16:11 crc kubenswrapper[4931]: I1129 00:16:11.249155 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-lc47s" podStartSLOduration=4.635274451 podStartE2EDuration="22.24911696s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:51.04143026 +0000 UTC m=+964.203323492" lastFinishedPulling="2025-11-29 00:16:08.655272729 +0000 UTC m=+981.817166001" observedRunningTime="2025-11-29 00:16:11.239880684 +0000 UTC m=+984.401773956" watchObservedRunningTime="2025-11-29 00:16:11.24911696 +0000 UTC m=+984.411023012" Nov 29 00:16:12 crc kubenswrapper[4931]: I1129 00:16:12.239801 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-lc47s" Nov 29 00:16:12 crc kubenswrapper[4931]: I1129 00:16:12.243647 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-lc47s" Nov 29 00:16:16 crc kubenswrapper[4931]: I1129 00:16:16.253725 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-74fb9455b7-sz2k9" Nov 29 00:16:16 crc kubenswrapper[4931]: I1129 00:16:16.295615 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f" event={"ID":"551c1c83-9dd7-47c7-8411-30881adc79bb","Type":"ContainerStarted","Data":"10d3616317332ea5a1949f1f4fb27721d3b94a7a46ff17ea9db08e14aa9c6c6b"} Nov 29 00:16:16 crc kubenswrapper[4931]: I1129 00:16:16.296698 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f" Nov 29 00:16:16 crc kubenswrapper[4931]: I1129 00:16:16.298630 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99" event={"ID":"c352f7a8-72dc-41da-9673-7fccd1a0974f","Type":"ContainerStarted","Data":"bcd9e961695a7991fc61b4bbd31c26a8543a62485bee54134154d1f57609f135"} Nov 29 00:16:16 crc kubenswrapper[4931]: I1129 00:16:16.299113 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99" Nov 29 00:16:16 crc kubenswrapper[4931]: I1129 00:16:16.301009 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99" Nov 29 00:16:16 crc kubenswrapper[4931]: I1129 00:16:16.302507 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f" Nov 29 00:16:16 crc kubenswrapper[4931]: I1129 00:16:16.327263 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tqw4f" podStartSLOduration=9.38172802 podStartE2EDuration="27.327238438s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:50.772907949 +0000 UTC m=+963.934801181" lastFinishedPulling="2025-11-29 00:16:08.718418367 +0000 UTC m=+981.880311599" observedRunningTime="2025-11-29 00:16:16.318383063 +0000 UTC m=+989.480276315" watchObservedRunningTime="2025-11-29 00:16:16.327238438 +0000 UTC m=+989.489131670" Nov 29 00:16:16 crc kubenswrapper[4931]: I1129 00:16:16.348109 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-lpf99" podStartSLOduration=9.72527219 podStartE2EDuration="27.348091649s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:51.032437609 +0000 UTC m=+964.194330841" lastFinishedPulling="2025-11-29 00:16:08.655257058 +0000 UTC m=+981.817150300" observedRunningTime="2025-11-29 00:16:16.339029228 +0000 UTC m=+989.500922480" watchObservedRunningTime="2025-11-29 00:16:16.348091649 +0000 UTC m=+989.509984881" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.307680 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st" event={"ID":"0ce6e577-25c0-4aee-b14f-47e02ebfbef9","Type":"ContainerStarted","Data":"4a0aec1d2a62abc68708af7f1ae19ffe9423e0a5024fc17306ce2be409bd2c4e"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.308004 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.309877 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.309961 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" event={"ID":"724defda-f0a4-4edd-8a1a-7f6625893fe7","Type":"ContainerStarted","Data":"479cad4720c2e2153317ba2a62cc246c9a4cb962aa9e35d082ce0b47f0210c0c"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.311570 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" event={"ID":"fd6606d5-8cd5-4308-a010-eea75cb666f1","Type":"ContainerStarted","Data":"380b47ac35840b7c1fa4562603d8e45736517e087082296f6c61ca10970cf716"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.311595 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" event={"ID":"fd6606d5-8cd5-4308-a010-eea75cb666f1","Type":"ContainerStarted","Data":"d581481d77e02feab8d146d779b1af2b6d2819a08b1b277550ba1b7693c5d576"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.311984 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.313678 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" event={"ID":"b57413c5-3888-496c-ad0b-18128019b945","Type":"ContainerStarted","Data":"e252a14aa90280e6915a61ae93a142a8abdd53d7559f355bdf5ce9a9fd907a89"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.313740 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" event={"ID":"b57413c5-3888-496c-ad0b-18128019b945","Type":"ContainerStarted","Data":"6a0905805fa83ee7796489c065df4b6f1a96d653de6784b46f31a7457e55349a"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.313947 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.316698 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" event={"ID":"5a12e787-bf00-4f2b-8760-9b9a5cad0834","Type":"ContainerStarted","Data":"1a550c72de610a4237c43018df4558a9870c05d7257fd80576e35a96fac00f21"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.318044 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4" event={"ID":"22500bd3-4958-4ba1-9b59-deb3032aea18","Type":"ContainerStarted","Data":"a1d910c662a398cad676ebc577149b982c0f018aa5b360aeb6825fa722e3061d"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.319640 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" event={"ID":"249844ae-a64c-4f9d-b204-bf3a1956aaa7","Type":"ContainerStarted","Data":"2d8dc2512e67b02908c873e7c916dbc22cda0b85109d0e078e415e8e76434dfb"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.319969 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.321637 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" event={"ID":"3a157886-d4b5-4019-a0d6-0b40f6509b21","Type":"ContainerStarted","Data":"40ac03ecee7c2dfc5c252c037da10b64b14b1e7a68f67b762a7b7261ad4ae5ef"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.321670 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" event={"ID":"3a157886-d4b5-4019-a0d6-0b40f6509b21","Type":"ContainerStarted","Data":"139b2f073c8893ac5580b24a04aef80fe9a36a0c65b10fdda72a14244bffcec5"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.321755 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.322225 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.323885 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7" event={"ID":"21842304-a783-4f75-bbb5-2ece9dacb6ca","Type":"ContainerStarted","Data":"9871d81689bea1aaf64c91dce609777c19e84e303b29a873f5f50a7be2820107"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.324110 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.325665 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" event={"ID":"6f4cd8c8-caf0-4191-a3c4-51a172812b9b","Type":"ContainerStarted","Data":"f6ec34b61fa141c59e3274ffd02bb9419b39f63ada03770ba1f620e3b3604bf9"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.326487 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.326525 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.328999 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" event={"ID":"9621ec40-d523-423e-9bbe-4b063355cf08","Type":"ContainerStarted","Data":"616f7b75b8137be98410debfe7df3921c84f5ebd053e5bf8d99924b4a06df9f5"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.329045 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.329256 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.330776 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" event={"ID":"2792db7c-cdd1-4356-97d5-fdba06bd44e4","Type":"ContainerStarted","Data":"9ac425b39d8fc66b089d9a6b6a9e7aa53f7269e521a5e70d4956b6e254a9637d"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.330824 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.330837 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" event={"ID":"2792db7c-cdd1-4356-97d5-fdba06bd44e4","Type":"ContainerStarted","Data":"d627495c1dc3f56675fb13261d3e41d70ce28be87d5ebf324b6a096984feb287"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.330984 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.332446 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" event={"ID":"365eb479-0132-4a0c-99d3-01cb923a9d61","Type":"ContainerStarted","Data":"b483680e0bf883272f35ce95ecaa351ea096c7b11d8d3a59379e510e2de537cd"} Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.338325 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-g82st" podStartSLOduration=10.44585578 podStartE2EDuration="28.338308549s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:50.846788227 +0000 UTC m=+964.008681459" lastFinishedPulling="2025-11-29 00:16:08.739240996 +0000 UTC m=+981.901134228" observedRunningTime="2025-11-29 00:16:17.334740876 +0000 UTC m=+990.496634348" watchObservedRunningTime="2025-11-29 00:16:17.338308549 +0000 UTC m=+990.500201781" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.352729 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-b4skl" podStartSLOduration=17.213290619 podStartE2EDuration="28.352708124s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:50.890613971 +0000 UTC m=+964.052507203" lastFinishedPulling="2025-11-29 00:16:02.030031466 +0000 UTC m=+975.191924708" observedRunningTime="2025-11-29 00:16:17.349860432 +0000 UTC m=+990.511753684" watchObservedRunningTime="2025-11-29 00:16:17.352708124 +0000 UTC m=+990.514601356" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.372355 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" podStartSLOduration=3.228081051 podStartE2EDuration="28.372328209s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:51.088698005 +0000 UTC m=+964.250591237" lastFinishedPulling="2025-11-29 00:16:16.232945143 +0000 UTC m=+989.394838395" observedRunningTime="2025-11-29 00:16:17.363080972 +0000 UTC m=+990.524974204" watchObservedRunningTime="2025-11-29 00:16:17.372328209 +0000 UTC m=+990.534221441" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.394144 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-nmtz7" podStartSLOduration=10.459592781 podStartE2EDuration="28.394126026s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:50.784254192 +0000 UTC m=+963.946147424" lastFinishedPulling="2025-11-29 00:16:08.718787437 +0000 UTC m=+981.880680669" observedRunningTime="2025-11-29 00:16:17.391192842 +0000 UTC m=+990.553086084" watchObservedRunningTime="2025-11-29 00:16:17.394126026 +0000 UTC m=+990.556019258" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.445623 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-4qjb4" podStartSLOduration=2.401874994 podStartE2EDuration="27.445608678s" podCreationTimestamp="2025-11-29 00:15:50 +0000 UTC" firstStartedPulling="2025-11-29 00:15:51.267769401 +0000 UTC m=+964.429662633" lastFinishedPulling="2025-11-29 00:16:16.311503085 +0000 UTC m=+989.473396317" observedRunningTime="2025-11-29 00:16:17.441522191 +0000 UTC m=+990.603415443" watchObservedRunningTime="2025-11-29 00:16:17.445608678 +0000 UTC m=+990.607501910" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.467925 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" podStartSLOduration=19.084785784 podStartE2EDuration="28.46789229s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:16:06.848106097 +0000 UTC m=+980.009999329" lastFinishedPulling="2025-11-29 00:16:16.231212573 +0000 UTC m=+989.393105835" observedRunningTime="2025-11-29 00:16:17.460881038 +0000 UTC m=+990.622774290" watchObservedRunningTime="2025-11-29 00:16:17.46789229 +0000 UTC m=+990.629785522" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.487373 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-dcl69" podStartSLOduration=10.999778624 podStartE2EDuration="28.48735682s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:51.25166495 +0000 UTC m=+964.413558182" lastFinishedPulling="2025-11-29 00:16:08.739243146 +0000 UTC m=+981.901136378" observedRunningTime="2025-11-29 00:16:17.484762176 +0000 UTC m=+990.646655418" watchObservedRunningTime="2025-11-29 00:16:17.48735682 +0000 UTC m=+990.649250052" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.514412 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-f62b4" podStartSLOduration=10.933055785 podStartE2EDuration="28.514394258s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:51.140630306 +0000 UTC m=+964.302523548" lastFinishedPulling="2025-11-29 00:16:08.721968789 +0000 UTC m=+981.883862021" observedRunningTime="2025-11-29 00:16:17.514393358 +0000 UTC m=+990.676286610" watchObservedRunningTime="2025-11-29 00:16:17.514394258 +0000 UTC m=+990.676287490" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.538458 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-q7q7q" podStartSLOduration=17.64405273 podStartE2EDuration="28.53844305s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:51.124221087 +0000 UTC m=+964.286114319" lastFinishedPulling="2025-11-29 00:16:02.018611397 +0000 UTC m=+975.180504639" observedRunningTime="2025-11-29 00:16:17.536191566 +0000 UTC m=+990.698084798" watchObservedRunningTime="2025-11-29 00:16:17.53844305 +0000 UTC m=+990.700336282" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.559573 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-2dbgl" podStartSLOduration=17.547176265 podStartE2EDuration="28.559557368s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:51.007677116 +0000 UTC m=+964.169570348" lastFinishedPulling="2025-11-29 00:16:02.020058209 +0000 UTC m=+975.181951451" observedRunningTime="2025-11-29 00:16:17.554488392 +0000 UTC m=+990.716381634" watchObservedRunningTime="2025-11-29 00:16:17.559557368 +0000 UTC m=+990.721450600" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.571296 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lnrrq" podStartSLOduration=11.047487125 podStartE2EDuration="28.571282256s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:51.132865918 +0000 UTC m=+964.294759150" lastFinishedPulling="2025-11-29 00:16:08.656660999 +0000 UTC m=+981.818554281" observedRunningTime="2025-11-29 00:16:17.56657682 +0000 UTC m=+990.728470052" watchObservedRunningTime="2025-11-29 00:16:17.571282256 +0000 UTC m=+990.733175488" Nov 29 00:16:17 crc kubenswrapper[4931]: I1129 00:16:17.588629 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" podStartSLOduration=3.612356042 podStartE2EDuration="28.588613975s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:15:51.258945515 +0000 UTC m=+964.420838747" lastFinishedPulling="2025-11-29 00:16:16.235203428 +0000 UTC m=+989.397096680" observedRunningTime="2025-11-29 00:16:17.585172856 +0000 UTC m=+990.747066098" watchObservedRunningTime="2025-11-29 00:16:17.588613975 +0000 UTC m=+990.750507207" Nov 29 00:16:23 crc kubenswrapper[4931]: I1129 00:16:23.735476 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:16:23 crc kubenswrapper[4931]: I1129 00:16:23.737181 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:16:25 crc kubenswrapper[4931]: I1129 00:16:25.515091 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nkmhr" Nov 29 00:16:25 crc kubenswrapper[4931]: I1129 00:16:25.546381 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" podStartSLOduration=27.756462584 podStartE2EDuration="36.54633245s" podCreationTimestamp="2025-11-29 00:15:49 +0000 UTC" firstStartedPulling="2025-11-29 00:16:07.447175245 +0000 UTC m=+980.609068477" lastFinishedPulling="2025-11-29 00:16:16.237045081 +0000 UTC m=+989.398938343" observedRunningTime="2025-11-29 00:16:17.620866733 +0000 UTC m=+990.782759975" watchObservedRunningTime="2025-11-29 00:16:25.54633245 +0000 UTC m=+998.708225692" Nov 29 00:16:25 crc kubenswrapper[4931]: I1129 00:16:25.749038 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp" Nov 29 00:16:30 crc kubenswrapper[4931]: I1129 00:16:30.371721 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-lthd8" Nov 29 00:16:30 crc kubenswrapper[4931]: I1129 00:16:30.603550 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-qp9ht" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.224296 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tghf9"] Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.226764 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.230029 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.230362 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.230496 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.230621 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-gknjz" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.231048 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tghf9"] Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.279545 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-jdhdt"] Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.280595 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.283087 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.312730 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-jdhdt"] Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.406193 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2-config\") pod \"dnsmasq-dns-675f4bcbfc-tghf9\" (UID: \"7cbb5cb5-d10a-4ee0-9345-fc5b651999d2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.406292 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/866218de-4228-42aa-a27e-3da9e742a615-config\") pod \"dnsmasq-dns-78dd6ddcc-jdhdt\" (UID: \"866218de-4228-42aa-a27e-3da9e742a615\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.406357 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj29g\" (UniqueName: \"kubernetes.io/projected/866218de-4228-42aa-a27e-3da9e742a615-kube-api-access-xj29g\") pod \"dnsmasq-dns-78dd6ddcc-jdhdt\" (UID: \"866218de-4228-42aa-a27e-3da9e742a615\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.406414 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x59sk\" (UniqueName: \"kubernetes.io/projected/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2-kube-api-access-x59sk\") pod \"dnsmasq-dns-675f4bcbfc-tghf9\" (UID: \"7cbb5cb5-d10a-4ee0-9345-fc5b651999d2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.406455 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/866218de-4228-42aa-a27e-3da9e742a615-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-jdhdt\" (UID: \"866218de-4228-42aa-a27e-3da9e742a615\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.507505 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/866218de-4228-42aa-a27e-3da9e742a615-config\") pod \"dnsmasq-dns-78dd6ddcc-jdhdt\" (UID: \"866218de-4228-42aa-a27e-3da9e742a615\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.507577 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj29g\" (UniqueName: \"kubernetes.io/projected/866218de-4228-42aa-a27e-3da9e742a615-kube-api-access-xj29g\") pod \"dnsmasq-dns-78dd6ddcc-jdhdt\" (UID: \"866218de-4228-42aa-a27e-3da9e742a615\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.507650 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x59sk\" (UniqueName: \"kubernetes.io/projected/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2-kube-api-access-x59sk\") pod \"dnsmasq-dns-675f4bcbfc-tghf9\" (UID: \"7cbb5cb5-d10a-4ee0-9345-fc5b651999d2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.507685 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/866218de-4228-42aa-a27e-3da9e742a615-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-jdhdt\" (UID: \"866218de-4228-42aa-a27e-3da9e742a615\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.507766 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2-config\") pod \"dnsmasq-dns-675f4bcbfc-tghf9\" (UID: \"7cbb5cb5-d10a-4ee0-9345-fc5b651999d2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.508297 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/866218de-4228-42aa-a27e-3da9e742a615-config\") pod \"dnsmasq-dns-78dd6ddcc-jdhdt\" (UID: \"866218de-4228-42aa-a27e-3da9e742a615\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.508686 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/866218de-4228-42aa-a27e-3da9e742a615-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-jdhdt\" (UID: \"866218de-4228-42aa-a27e-3da9e742a615\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.508898 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2-config\") pod \"dnsmasq-dns-675f4bcbfc-tghf9\" (UID: \"7cbb5cb5-d10a-4ee0-9345-fc5b651999d2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.528369 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj29g\" (UniqueName: \"kubernetes.io/projected/866218de-4228-42aa-a27e-3da9e742a615-kube-api-access-xj29g\") pod \"dnsmasq-dns-78dd6ddcc-jdhdt\" (UID: \"866218de-4228-42aa-a27e-3da9e742a615\") " pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.532943 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x59sk\" (UniqueName: \"kubernetes.io/projected/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2-kube-api-access-x59sk\") pod \"dnsmasq-dns-675f4bcbfc-tghf9\" (UID: \"7cbb5cb5-d10a-4ee0-9345-fc5b651999d2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.544217 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.596730 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.735074 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:16:53 crc kubenswrapper[4931]: I1129 00:16:53.735146 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:16:54 crc kubenswrapper[4931]: I1129 00:16:54.047333 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tghf9"] Nov 29 00:16:54 crc kubenswrapper[4931]: I1129 00:16:54.053632 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 00:16:54 crc kubenswrapper[4931]: I1129 00:16:54.158532 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-jdhdt"] Nov 29 00:16:54 crc kubenswrapper[4931]: W1129 00:16:54.164119 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod866218de_4228_42aa_a27e_3da9e742a615.slice/crio-50740c3eeef2b323d0df957485fe4141eac9f4c0dfe2a97b36710cea91f9cb0c WatchSource:0}: Error finding container 50740c3eeef2b323d0df957485fe4141eac9f4c0dfe2a97b36710cea91f9cb0c: Status 404 returned error can't find the container with id 50740c3eeef2b323d0df957485fe4141eac9f4c0dfe2a97b36710cea91f9cb0c Nov 29 00:16:54 crc kubenswrapper[4931]: I1129 00:16:54.683369 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" event={"ID":"866218de-4228-42aa-a27e-3da9e742a615","Type":"ContainerStarted","Data":"50740c3eeef2b323d0df957485fe4141eac9f4c0dfe2a97b36710cea91f9cb0c"} Nov 29 00:16:54 crc kubenswrapper[4931]: I1129 00:16:54.685399 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" event={"ID":"7cbb5cb5-d10a-4ee0-9345-fc5b651999d2","Type":"ContainerStarted","Data":"06611ae94cbd1e0ed38277714441981dd3c3129a35aa10d5027019cef5b8cb24"} Nov 29 00:16:55 crc kubenswrapper[4931]: I1129 00:16:55.901868 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tghf9"] Nov 29 00:16:55 crc kubenswrapper[4931]: I1129 00:16:55.905880 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-8mx5k"] Nov 29 00:16:55 crc kubenswrapper[4931]: I1129 00:16:55.911734 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:16:55 crc kubenswrapper[4931]: I1129 00:16:55.918282 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-8mx5k"] Nov 29 00:16:55 crc kubenswrapper[4931]: I1129 00:16:55.942191 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbbxx\" (UniqueName: \"kubernetes.io/projected/c3a0e1d0-d484-4b42-81a1-f0520ef32717-kube-api-access-jbbxx\") pod \"dnsmasq-dns-666b6646f7-8mx5k\" (UID: \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\") " pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:16:55 crc kubenswrapper[4931]: I1129 00:16:55.942231 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a0e1d0-d484-4b42-81a1-f0520ef32717-config\") pod \"dnsmasq-dns-666b6646f7-8mx5k\" (UID: \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\") " pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:16:55 crc kubenswrapper[4931]: I1129 00:16:55.942268 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3a0e1d0-d484-4b42-81a1-f0520ef32717-dns-svc\") pod \"dnsmasq-dns-666b6646f7-8mx5k\" (UID: \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\") " pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.044826 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbbxx\" (UniqueName: \"kubernetes.io/projected/c3a0e1d0-d484-4b42-81a1-f0520ef32717-kube-api-access-jbbxx\") pod \"dnsmasq-dns-666b6646f7-8mx5k\" (UID: \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\") " pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.044875 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a0e1d0-d484-4b42-81a1-f0520ef32717-config\") pod \"dnsmasq-dns-666b6646f7-8mx5k\" (UID: \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\") " pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.044920 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3a0e1d0-d484-4b42-81a1-f0520ef32717-dns-svc\") pod \"dnsmasq-dns-666b6646f7-8mx5k\" (UID: \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\") " pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.046007 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3a0e1d0-d484-4b42-81a1-f0520ef32717-dns-svc\") pod \"dnsmasq-dns-666b6646f7-8mx5k\" (UID: \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\") " pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.046170 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a0e1d0-d484-4b42-81a1-f0520ef32717-config\") pod \"dnsmasq-dns-666b6646f7-8mx5k\" (UID: \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\") " pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.069911 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbbxx\" (UniqueName: \"kubernetes.io/projected/c3a0e1d0-d484-4b42-81a1-f0520ef32717-kube-api-access-jbbxx\") pod \"dnsmasq-dns-666b6646f7-8mx5k\" (UID: \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\") " pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.135997 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-jdhdt"] Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.154025 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8jzz9"] Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.157156 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.166641 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8jzz9"] Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.239483 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.350322 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a0b54682-21aa-4382-b7c4-5db01da1cab8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8jzz9\" (UID: \"a0b54682-21aa-4382-b7c4-5db01da1cab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.350392 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr2pk\" (UniqueName: \"kubernetes.io/projected/a0b54682-21aa-4382-b7c4-5db01da1cab8-kube-api-access-mr2pk\") pod \"dnsmasq-dns-57d769cc4f-8jzz9\" (UID: \"a0b54682-21aa-4382-b7c4-5db01da1cab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.350425 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b54682-21aa-4382-b7c4-5db01da1cab8-config\") pod \"dnsmasq-dns-57d769cc4f-8jzz9\" (UID: \"a0b54682-21aa-4382-b7c4-5db01da1cab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.451430 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a0b54682-21aa-4382-b7c4-5db01da1cab8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8jzz9\" (UID: \"a0b54682-21aa-4382-b7c4-5db01da1cab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.451508 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr2pk\" (UniqueName: \"kubernetes.io/projected/a0b54682-21aa-4382-b7c4-5db01da1cab8-kube-api-access-mr2pk\") pod \"dnsmasq-dns-57d769cc4f-8jzz9\" (UID: \"a0b54682-21aa-4382-b7c4-5db01da1cab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.451537 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b54682-21aa-4382-b7c4-5db01da1cab8-config\") pod \"dnsmasq-dns-57d769cc4f-8jzz9\" (UID: \"a0b54682-21aa-4382-b7c4-5db01da1cab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.452399 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b54682-21aa-4382-b7c4-5db01da1cab8-config\") pod \"dnsmasq-dns-57d769cc4f-8jzz9\" (UID: \"a0b54682-21aa-4382-b7c4-5db01da1cab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.452647 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a0b54682-21aa-4382-b7c4-5db01da1cab8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8jzz9\" (UID: \"a0b54682-21aa-4382-b7c4-5db01da1cab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.469442 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr2pk\" (UniqueName: \"kubernetes.io/projected/a0b54682-21aa-4382-b7c4-5db01da1cab8-kube-api-access-mr2pk\") pod \"dnsmasq-dns-57d769cc4f-8jzz9\" (UID: \"a0b54682-21aa-4382-b7c4-5db01da1cab8\") " pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.492257 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.700484 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-8mx5k"] Nov 29 00:16:56 crc kubenswrapper[4931]: W1129 00:16:56.708624 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3a0e1d0_d484_4b42_81a1_f0520ef32717.slice/crio-32b787ad030bda7449cdc19ba198cc2aff751d0a4ceae79815f83c386531d60d WatchSource:0}: Error finding container 32b787ad030bda7449cdc19ba198cc2aff751d0a4ceae79815f83c386531d60d: Status 404 returned error can't find the container with id 32b787ad030bda7449cdc19ba198cc2aff751d0a4ceae79815f83c386531d60d Nov 29 00:16:56 crc kubenswrapper[4931]: I1129 00:16:56.898898 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8jzz9"] Nov 29 00:16:56 crc kubenswrapper[4931]: W1129 00:16:56.907731 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0b54682_21aa_4382_b7c4_5db01da1cab8.slice/crio-c95d509507dde8b4c99413c0a84d8a5f416aec93749f4f74154f8666b413ef17 WatchSource:0}: Error finding container c95d509507dde8b4c99413c0a84d8a5f416aec93749f4f74154f8666b413ef17: Status 404 returned error can't find the container with id c95d509507dde8b4c99413c0a84d8a5f416aec93749f4f74154f8666b413ef17 Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.044576 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.049762 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.051713 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.052044 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.052206 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.052337 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.052664 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.052844 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-h4vzq" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.052988 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.054231 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.161323 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.161641 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hhmg\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-kube-api-access-2hhmg\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.161666 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.161714 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.161737 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.161757 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.161790 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.161821 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.161852 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-config-data\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.161883 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0d661705-7bb5-482f-b2a8-79b44f6351ac-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.161903 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0d661705-7bb5-482f-b2a8-79b44f6351ac-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.263649 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.263714 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.263742 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.263785 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.263851 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.263934 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-config-data\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.263993 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0d661705-7bb5-482f-b2a8-79b44f6351ac-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.264020 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0d661705-7bb5-482f-b2a8-79b44f6351ac-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.264044 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.264062 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hhmg\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-kube-api-access-2hhmg\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.264083 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.264251 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.264285 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.264441 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.264699 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.266154 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.268879 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-config-data\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.271304 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0d661705-7bb5-482f-b2a8-79b44f6351ac-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.276044 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.277458 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0d661705-7bb5-482f-b2a8-79b44f6351ac-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.277721 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.283966 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hhmg\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-kube-api-access-2hhmg\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.291778 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.312141 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.313404 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.319530 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.320491 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.320596 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.320676 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.320691 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wz4dt" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.320707 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.320854 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.328095 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.398389 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.467049 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.467171 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fbd78afc-2d6d-4118-92a2-e3499560359f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.467217 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.467246 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.467270 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fbd78afc-2d6d-4118-92a2-e3499560359f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.467293 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.467371 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.467433 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.468187 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwjd9\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-kube-api-access-qwjd9\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.468227 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.469344 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.570565 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fbd78afc-2d6d-4118-92a2-e3499560359f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.570618 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.570638 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.570656 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fbd78afc-2d6d-4118-92a2-e3499560359f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.570673 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.570707 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.570723 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.570737 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwjd9\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-kube-api-access-qwjd9\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.570753 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.570781 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.570827 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.571395 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.571792 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.572508 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.572566 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.572615 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.572727 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.575077 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fbd78afc-2d6d-4118-92a2-e3499560359f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.575364 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fbd78afc-2d6d-4118-92a2-e3499560359f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.575425 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.577336 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.591551 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwjd9\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-kube-api-access-qwjd9\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.595912 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.648827 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.726898 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" event={"ID":"c3a0e1d0-d484-4b42-81a1-f0520ef32717","Type":"ContainerStarted","Data":"32b787ad030bda7449cdc19ba198cc2aff751d0a4ceae79815f83c386531d60d"} Nov 29 00:16:57 crc kubenswrapper[4931]: I1129 00:16:57.728357 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" event={"ID":"a0b54682-21aa-4382-b7c4-5db01da1cab8","Type":"ContainerStarted","Data":"c95d509507dde8b4c99413c0a84d8a5f416aec93749f4f74154f8666b413ef17"} Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.705839 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.709134 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.711291 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.712068 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-5m584" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.712432 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.719513 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.720442 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.722501 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.890994 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba2777f2-23db-46b8-8ccf-cb0813a39678-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.891057 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ba2777f2-23db-46b8-8ccf-cb0813a39678-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.891183 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba2777f2-23db-46b8-8ccf-cb0813a39678-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.891222 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba2777f2-23db-46b8-8ccf-cb0813a39678-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.891286 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb8qh\" (UniqueName: \"kubernetes.io/projected/ba2777f2-23db-46b8-8ccf-cb0813a39678-kube-api-access-jb8qh\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.891379 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ba2777f2-23db-46b8-8ccf-cb0813a39678-config-data-default\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.891485 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.891587 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ba2777f2-23db-46b8-8ccf-cb0813a39678-kolla-config\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.992656 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb8qh\" (UniqueName: \"kubernetes.io/projected/ba2777f2-23db-46b8-8ccf-cb0813a39678-kube-api-access-jb8qh\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.992747 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ba2777f2-23db-46b8-8ccf-cb0813a39678-config-data-default\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.992782 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.992865 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ba2777f2-23db-46b8-8ccf-cb0813a39678-kolla-config\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.992911 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba2777f2-23db-46b8-8ccf-cb0813a39678-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.992947 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ba2777f2-23db-46b8-8ccf-cb0813a39678-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.992973 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba2777f2-23db-46b8-8ccf-cb0813a39678-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.993002 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba2777f2-23db-46b8-8ccf-cb0813a39678-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.993265 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.993576 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ba2777f2-23db-46b8-8ccf-cb0813a39678-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.993846 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ba2777f2-23db-46b8-8ccf-cb0813a39678-kolla-config\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.993903 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ba2777f2-23db-46b8-8ccf-cb0813a39678-config-data-default\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.995496 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba2777f2-23db-46b8-8ccf-cb0813a39678-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.999527 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba2777f2-23db-46b8-8ccf-cb0813a39678-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:58 crc kubenswrapper[4931]: I1129 00:16:58.999601 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba2777f2-23db-46b8-8ccf-cb0813a39678-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:59 crc kubenswrapper[4931]: I1129 00:16:59.014751 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb8qh\" (UniqueName: \"kubernetes.io/projected/ba2777f2-23db-46b8-8ccf-cb0813a39678-kube-api-access-jb8qh\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:59 crc kubenswrapper[4931]: I1129 00:16:59.020841 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"ba2777f2-23db-46b8-8ccf-cb0813a39678\") " pod="openstack/openstack-galera-0" Nov 29 00:16:59 crc kubenswrapper[4931]: I1129 00:16:59.032949 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.257076 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.259232 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.264575 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.264799 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.265753 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-h6fxn" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.268617 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.274102 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.416600 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b4d256-9b86-48e3-98d8-266271a338dd-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.416878 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2b4d256-9b86-48e3-98d8-266271a338dd-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.416966 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2b4d256-9b86-48e3-98d8-266271a338dd-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.417202 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c2b4d256-9b86-48e3-98d8-266271a338dd-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.417265 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.417313 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2b4d256-9b86-48e3-98d8-266271a338dd-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.417395 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4twb\" (UniqueName: \"kubernetes.io/projected/c2b4d256-9b86-48e3-98d8-266271a338dd-kube-api-access-l4twb\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.417435 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c2b4d256-9b86-48e3-98d8-266271a338dd-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.519032 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c2b4d256-9b86-48e3-98d8-266271a338dd-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.519092 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.519118 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2b4d256-9b86-48e3-98d8-266271a338dd-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.519176 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4twb\" (UniqueName: \"kubernetes.io/projected/c2b4d256-9b86-48e3-98d8-266271a338dd-kube-api-access-l4twb\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.519208 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c2b4d256-9b86-48e3-98d8-266271a338dd-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.519239 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b4d256-9b86-48e3-98d8-266271a338dd-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.519280 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2b4d256-9b86-48e3-98d8-266271a338dd-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.519307 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2b4d256-9b86-48e3-98d8-266271a338dd-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.519565 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c2b4d256-9b86-48e3-98d8-266271a338dd-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.520097 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2b4d256-9b86-48e3-98d8-266271a338dd-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.520322 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c2b4d256-9b86-48e3-98d8-266271a338dd-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.520513 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.521269 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2b4d256-9b86-48e3-98d8-266271a338dd-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.525771 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b4d256-9b86-48e3-98d8-266271a338dd-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.542137 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2b4d256-9b86-48e3-98d8-266271a338dd-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.546721 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4twb\" (UniqueName: \"kubernetes.io/projected/c2b4d256-9b86-48e3-98d8-266271a338dd-kube-api-access-l4twb\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.567148 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c2b4d256-9b86-48e3-98d8-266271a338dd\") " pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.580223 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.609128 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.611452 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.615878 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.616265 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-dwz4s" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.616444 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.616450 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.733350 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d60ec023-b0d6-466c-b188-5e0e2b47a962-kolla-config\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.733430 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60ec023-b0d6-466c-b188-5e0e2b47a962-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.733577 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ldmw\" (UniqueName: \"kubernetes.io/projected/d60ec023-b0d6-466c-b188-5e0e2b47a962-kube-api-access-5ldmw\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.733637 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d60ec023-b0d6-466c-b188-5e0e2b47a962-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.733662 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d60ec023-b0d6-466c-b188-5e0e2b47a962-config-data\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.835410 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d60ec023-b0d6-466c-b188-5e0e2b47a962-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.835459 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d60ec023-b0d6-466c-b188-5e0e2b47a962-config-data\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.835482 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d60ec023-b0d6-466c-b188-5e0e2b47a962-kolla-config\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.835530 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60ec023-b0d6-466c-b188-5e0e2b47a962-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.835561 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ldmw\" (UniqueName: \"kubernetes.io/projected/d60ec023-b0d6-466c-b188-5e0e2b47a962-kube-api-access-5ldmw\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.836384 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d60ec023-b0d6-466c-b188-5e0e2b47a962-config-data\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.836453 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d60ec023-b0d6-466c-b188-5e0e2b47a962-kolla-config\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.842163 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d60ec023-b0d6-466c-b188-5e0e2b47a962-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.855280 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d60ec023-b0d6-466c-b188-5e0e2b47a962-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.856287 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ldmw\" (UniqueName: \"kubernetes.io/projected/d60ec023-b0d6-466c-b188-5e0e2b47a962-kube-api-access-5ldmw\") pod \"memcached-0\" (UID: \"d60ec023-b0d6-466c-b188-5e0e2b47a962\") " pod="openstack/memcached-0" Nov 29 00:17:00 crc kubenswrapper[4931]: I1129 00:17:00.957852 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 29 00:17:02 crc kubenswrapper[4931]: I1129 00:17:02.285442 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 00:17:02 crc kubenswrapper[4931]: I1129 00:17:02.287366 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 29 00:17:02 crc kubenswrapper[4931]: I1129 00:17:02.289701 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-krlxr" Nov 29 00:17:02 crc kubenswrapper[4931]: I1129 00:17:02.305882 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 00:17:02 crc kubenswrapper[4931]: I1129 00:17:02.461615 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsvwl\" (UniqueName: \"kubernetes.io/projected/be472357-f01e-4e83-8903-49c0e2f60c97-kube-api-access-qsvwl\") pod \"kube-state-metrics-0\" (UID: \"be472357-f01e-4e83-8903-49c0e2f60c97\") " pod="openstack/kube-state-metrics-0" Nov 29 00:17:02 crc kubenswrapper[4931]: I1129 00:17:02.563601 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsvwl\" (UniqueName: \"kubernetes.io/projected/be472357-f01e-4e83-8903-49c0e2f60c97-kube-api-access-qsvwl\") pod \"kube-state-metrics-0\" (UID: \"be472357-f01e-4e83-8903-49c0e2f60c97\") " pod="openstack/kube-state-metrics-0" Nov 29 00:17:02 crc kubenswrapper[4931]: I1129 00:17:02.595631 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsvwl\" (UniqueName: \"kubernetes.io/projected/be472357-f01e-4e83-8903-49c0e2f60c97-kube-api-access-qsvwl\") pod \"kube-state-metrics-0\" (UID: \"be472357-f01e-4e83-8903-49c0e2f60c97\") " pod="openstack/kube-state-metrics-0" Nov 29 00:17:02 crc kubenswrapper[4931]: I1129 00:17:02.602396 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.465772 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-trsph"] Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.467064 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.469235 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.469384 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-j5rlw" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.469396 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.491676 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-trsph"] Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.516060 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-7s9fx"] Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.525784 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.528781 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a64bc6bd-8033-45c9-b2d7-9602aea09225-var-run\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.532204 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64bc6bd-8033-45c9-b2d7-9602aea09225-combined-ca-bundle\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.532458 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a64bc6bd-8033-45c9-b2d7-9602aea09225-scripts\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.532574 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a64bc6bd-8033-45c9-b2d7-9602aea09225-var-log-ovn\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.532685 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljxgj\" (UniqueName: \"kubernetes.io/projected/a64bc6bd-8033-45c9-b2d7-9602aea09225-kube-api-access-ljxgj\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.532775 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a64bc6bd-8033-45c9-b2d7-9602aea09225-ovn-controller-tls-certs\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.532867 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a64bc6bd-8033-45c9-b2d7-9602aea09225-var-run-ovn\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.530856 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7s9fx"] Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.634706 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a64bc6bd-8033-45c9-b2d7-9602aea09225-ovn-controller-tls-certs\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.634769 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a64bc6bd-8033-45c9-b2d7-9602aea09225-var-run-ovn\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.634797 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/28c11b0b-3026-4f9f-8de8-71174fbd19d6-var-run\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.634857 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a64bc6bd-8033-45c9-b2d7-9602aea09225-var-run\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.634881 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxrfj\" (UniqueName: \"kubernetes.io/projected/28c11b0b-3026-4f9f-8de8-71174fbd19d6-kube-api-access-fxrfj\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.634935 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/28c11b0b-3026-4f9f-8de8-71174fbd19d6-etc-ovs\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.634954 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28c11b0b-3026-4f9f-8de8-71174fbd19d6-scripts\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.634993 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/28c11b0b-3026-4f9f-8de8-71174fbd19d6-var-lib\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.635014 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64bc6bd-8033-45c9-b2d7-9602aea09225-combined-ca-bundle\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.635034 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/28c11b0b-3026-4f9f-8de8-71174fbd19d6-var-log\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.635077 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a64bc6bd-8033-45c9-b2d7-9602aea09225-scripts\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.635121 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a64bc6bd-8033-45c9-b2d7-9602aea09225-var-log-ovn\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.635160 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljxgj\" (UniqueName: \"kubernetes.io/projected/a64bc6bd-8033-45c9-b2d7-9602aea09225-kube-api-access-ljxgj\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.636166 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a64bc6bd-8033-45c9-b2d7-9602aea09225-var-log-ovn\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.636239 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a64bc6bd-8033-45c9-b2d7-9602aea09225-var-run-ovn\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.636283 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a64bc6bd-8033-45c9-b2d7-9602aea09225-var-run\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.637935 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a64bc6bd-8033-45c9-b2d7-9602aea09225-scripts\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.640531 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64bc6bd-8033-45c9-b2d7-9602aea09225-combined-ca-bundle\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.650268 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a64bc6bd-8033-45c9-b2d7-9602aea09225-ovn-controller-tls-certs\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.654506 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljxgj\" (UniqueName: \"kubernetes.io/projected/a64bc6bd-8033-45c9-b2d7-9602aea09225-kube-api-access-ljxgj\") pod \"ovn-controller-trsph\" (UID: \"a64bc6bd-8033-45c9-b2d7-9602aea09225\") " pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.739311 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/28c11b0b-3026-4f9f-8de8-71174fbd19d6-var-run\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.739388 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxrfj\" (UniqueName: \"kubernetes.io/projected/28c11b0b-3026-4f9f-8de8-71174fbd19d6-kube-api-access-fxrfj\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.739450 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/28c11b0b-3026-4f9f-8de8-71174fbd19d6-etc-ovs\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.739465 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28c11b0b-3026-4f9f-8de8-71174fbd19d6-scripts\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.739470 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/28c11b0b-3026-4f9f-8de8-71174fbd19d6-var-run\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.739524 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/28c11b0b-3026-4f9f-8de8-71174fbd19d6-var-lib\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.739544 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/28c11b0b-3026-4f9f-8de8-71174fbd19d6-var-log\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.739718 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/28c11b0b-3026-4f9f-8de8-71174fbd19d6-var-lib\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.739772 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/28c11b0b-3026-4f9f-8de8-71174fbd19d6-var-log\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.739895 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/28c11b0b-3026-4f9f-8de8-71174fbd19d6-etc-ovs\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.744164 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28c11b0b-3026-4f9f-8de8-71174fbd19d6-scripts\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.756316 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxrfj\" (UniqueName: \"kubernetes.io/projected/28c11b0b-3026-4f9f-8de8-71174fbd19d6-kube-api-access-fxrfj\") pod \"ovn-controller-ovs-7s9fx\" (UID: \"28c11b0b-3026-4f9f-8de8-71174fbd19d6\") " pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.789177 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-trsph" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.843587 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.922464 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.923846 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.926055 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.926090 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.927006 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-jtklj" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.927522 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.927578 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 29 00:17:06 crc kubenswrapper[4931]: I1129 00:17:06.932459 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.048579 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/979903af-0c72-45fe-a3de-0dc52370bc57-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.048967 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcmmx\" (UniqueName: \"kubernetes.io/projected/979903af-0c72-45fe-a3de-0dc52370bc57-kube-api-access-lcmmx\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.049013 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/979903af-0c72-45fe-a3de-0dc52370bc57-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.049088 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.049109 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/979903af-0c72-45fe-a3de-0dc52370bc57-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.049151 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/979903af-0c72-45fe-a3de-0dc52370bc57-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.049176 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/979903af-0c72-45fe-a3de-0dc52370bc57-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.049224 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/979903af-0c72-45fe-a3de-0dc52370bc57-config\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.151164 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.151209 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/979903af-0c72-45fe-a3de-0dc52370bc57-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.151251 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/979903af-0c72-45fe-a3de-0dc52370bc57-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.151270 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/979903af-0c72-45fe-a3de-0dc52370bc57-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.151310 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/979903af-0c72-45fe-a3de-0dc52370bc57-config\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.151341 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/979903af-0c72-45fe-a3de-0dc52370bc57-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.151359 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcmmx\" (UniqueName: \"kubernetes.io/projected/979903af-0c72-45fe-a3de-0dc52370bc57-kube-api-access-lcmmx\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.151379 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/979903af-0c72-45fe-a3de-0dc52370bc57-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.151471 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.152040 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/979903af-0c72-45fe-a3de-0dc52370bc57-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.152510 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/979903af-0c72-45fe-a3de-0dc52370bc57-config\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.152791 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/979903af-0c72-45fe-a3de-0dc52370bc57-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.155717 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/979903af-0c72-45fe-a3de-0dc52370bc57-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.159004 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/979903af-0c72-45fe-a3de-0dc52370bc57-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.167669 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcmmx\" (UniqueName: \"kubernetes.io/projected/979903af-0c72-45fe-a3de-0dc52370bc57-kube-api-access-lcmmx\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.174245 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.176193 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/979903af-0c72-45fe-a3de-0dc52370bc57-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"979903af-0c72-45fe-a3de-0dc52370bc57\") " pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:07 crc kubenswrapper[4931]: I1129 00:17:07.248570 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.561368 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.563055 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.566805 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-bcnhm" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.566928 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.573341 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.573491 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.573794 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.695567 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37afe9f6-664f-4f51-90a9-485e87a1db92-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.695656 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37afe9f6-664f-4f51-90a9-485e87a1db92-config\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.695754 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/37afe9f6-664f-4f51-90a9-485e87a1db92-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.695814 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37afe9f6-664f-4f51-90a9-485e87a1db92-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.695948 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37afe9f6-664f-4f51-90a9-485e87a1db92-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.696040 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/37afe9f6-664f-4f51-90a9-485e87a1db92-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.696132 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.696197 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnbrt\" (UniqueName: \"kubernetes.io/projected/37afe9f6-664f-4f51-90a9-485e87a1db92-kube-api-access-hnbrt\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.797551 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37afe9f6-664f-4f51-90a9-485e87a1db92-config\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.797613 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/37afe9f6-664f-4f51-90a9-485e87a1db92-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.797638 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37afe9f6-664f-4f51-90a9-485e87a1db92-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.797694 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37afe9f6-664f-4f51-90a9-485e87a1db92-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.797734 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/37afe9f6-664f-4f51-90a9-485e87a1db92-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.797755 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.797784 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnbrt\" (UniqueName: \"kubernetes.io/projected/37afe9f6-664f-4f51-90a9-485e87a1db92-kube-api-access-hnbrt\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.797884 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37afe9f6-664f-4f51-90a9-485e87a1db92-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.798983 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.799086 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37afe9f6-664f-4f51-90a9-485e87a1db92-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.799514 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37afe9f6-664f-4f51-90a9-485e87a1db92-config\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.799997 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/37afe9f6-664f-4f51-90a9-485e87a1db92-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.806561 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37afe9f6-664f-4f51-90a9-485e87a1db92-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.806796 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/37afe9f6-664f-4f51-90a9-485e87a1db92-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.815365 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37afe9f6-664f-4f51-90a9-485e87a1db92-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.831047 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.836452 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnbrt\" (UniqueName: \"kubernetes.io/projected/37afe9f6-664f-4f51-90a9-485e87a1db92-kube-api-access-hnbrt\") pod \"ovsdbserver-sb-0\" (UID: \"37afe9f6-664f-4f51-90a9-485e87a1db92\") " pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:09 crc kubenswrapper[4931]: I1129 00:17:09.897686 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:12 crc kubenswrapper[4931]: E1129 00:17:12.993775 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 29 00:17:12 crc kubenswrapper[4931]: E1129 00:17:12.994489 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jbbxx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-8mx5k_openstack(c3a0e1d0-d484-4b42-81a1-f0520ef32717): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 00:17:12 crc kubenswrapper[4931]: E1129 00:17:12.996001 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" podUID="c3a0e1d0-d484-4b42-81a1-f0520ef32717" Nov 29 00:17:13 crc kubenswrapper[4931]: E1129 00:17:13.012301 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 29 00:17:13 crc kubenswrapper[4931]: E1129 00:17:13.012489 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xj29g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-jdhdt_openstack(866218de-4228-42aa-a27e-3da9e742a615): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 00:17:13 crc kubenswrapper[4931]: E1129 00:17:13.013717 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" podUID="866218de-4228-42aa-a27e-3da9e742a615" Nov 29 00:17:13 crc kubenswrapper[4931]: E1129 00:17:13.139121 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 29 00:17:13 crc kubenswrapper[4931]: E1129 00:17:13.139486 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x59sk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-tghf9_openstack(7cbb5cb5-d10a-4ee0-9345-fc5b651999d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 00:17:13 crc kubenswrapper[4931]: E1129 00:17:13.141185 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" podUID="7cbb5cb5-d10a-4ee0-9345-fc5b651999d2" Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.424508 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 00:17:13 crc kubenswrapper[4931]: W1129 00:17:13.426419 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d661705_7bb5_482f_b2a8_79b44f6351ac.slice/crio-4e5115ada1c17ffa4b9282df255656354f160bb6c4761063ca2718a5f5294d12 WatchSource:0}: Error finding container 4e5115ada1c17ffa4b9282df255656354f160bb6c4761063ca2718a5f5294d12: Status 404 returned error can't find the container with id 4e5115ada1c17ffa4b9282df255656354f160bb6c4761063ca2718a5f5294d12 Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.653828 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 29 00:17:13 crc kubenswrapper[4931]: W1129 00:17:13.672553 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2b4d256_9b86_48e3_98d8_266271a338dd.slice/crio-6d73e7a815fdb291e0d737d661a93b02398abef8451fd5e60bf2de577368e343 WatchSource:0}: Error finding container 6d73e7a815fdb291e0d737d661a93b02398abef8451fd5e60bf2de577368e343: Status 404 returned error can't find the container with id 6d73e7a815fdb291e0d737d661a93b02398abef8451fd5e60bf2de577368e343 Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.677554 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 29 00:17:13 crc kubenswrapper[4931]: W1129 00:17:13.694136 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbd78afc_2d6d_4118_92a2_e3499560359f.slice/crio-d17cdeaa13e0bb877cdb7326736d7f10f567d69ed0d7fcdba99c86f897e3bc3c WatchSource:0}: Error finding container d17cdeaa13e0bb877cdb7326736d7f10f567d69ed0d7fcdba99c86f897e3bc3c: Status 404 returned error can't find the container with id d17cdeaa13e0bb877cdb7326736d7f10f567d69ed0d7fcdba99c86f897e3bc3c Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.695219 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.801978 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7s9fx"] Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.813124 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-trsph"] Nov 29 00:17:13 crc kubenswrapper[4931]: W1129 00:17:13.827189 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd60ec023_b0d6_466c_b188_5e0e2b47a962.slice/crio-b3e5a4ddd08f49bc9a147cade5766a54e965f698904b2bdbb6e49c14fa63de06 WatchSource:0}: Error finding container b3e5a4ddd08f49bc9a147cade5766a54e965f698904b2bdbb6e49c14fa63de06: Status 404 returned error can't find the container with id b3e5a4ddd08f49bc9a147cade5766a54e965f698904b2bdbb6e49c14fa63de06 Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.827709 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.893883 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 29 00:17:13 crc kubenswrapper[4931]: W1129 00:17:13.894977 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37afe9f6_664f_4f51_90a9_485e87a1db92.slice/crio-ed054c8d65caea3189e1fc415eba8601c54c23c7fda29ba3ec2056fa6d28ce28 WatchSource:0}: Error finding container ed054c8d65caea3189e1fc415eba8601c54c23c7fda29ba3ec2056fa6d28ce28: Status 404 returned error can't find the container with id ed054c8d65caea3189e1fc415eba8601c54c23c7fda29ba3ec2056fa6d28ce28 Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.913183 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c2b4d256-9b86-48e3-98d8-266271a338dd","Type":"ContainerStarted","Data":"6d73e7a815fdb291e0d737d661a93b02398abef8451fd5e60bf2de577368e343"} Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.914335 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"37afe9f6-664f-4f51-90a9-485e87a1db92","Type":"ContainerStarted","Data":"ed054c8d65caea3189e1fc415eba8601c54c23c7fda29ba3ec2056fa6d28ce28"} Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.915899 4931 generic.go:334] "Generic (PLEG): container finished" podID="a0b54682-21aa-4382-b7c4-5db01da1cab8" containerID="21e5487bc5744a60090acfcddcd06012d736d258de0c9c771c2aac432a6910e8" exitCode=0 Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.915950 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" event={"ID":"a0b54682-21aa-4382-b7c4-5db01da1cab8","Type":"ContainerDied","Data":"21e5487bc5744a60090acfcddcd06012d736d258de0c9c771c2aac432a6910e8"} Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.916752 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-trsph" event={"ID":"a64bc6bd-8033-45c9-b2d7-9602aea09225","Type":"ContainerStarted","Data":"262b4250624c0af30ba180a15ce803f2eb0678411d42259db6ceb0b2fac4eebe"} Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.919548 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d60ec023-b0d6-466c-b188-5e0e2b47a962","Type":"ContainerStarted","Data":"b3e5a4ddd08f49bc9a147cade5766a54e965f698904b2bdbb6e49c14fa63de06"} Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.920369 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fbd78afc-2d6d-4118-92a2-e3499560359f","Type":"ContainerStarted","Data":"d17cdeaa13e0bb877cdb7326736d7f10f567d69ed0d7fcdba99c86f897e3bc3c"} Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.921119 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ba2777f2-23db-46b8-8ccf-cb0813a39678","Type":"ContainerStarted","Data":"d72452a8bbd5a15be61b9b1cc44552e7f91749778fcc15691b1848b468a2f6f2"} Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.921953 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7s9fx" event={"ID":"28c11b0b-3026-4f9f-8de8-71174fbd19d6","Type":"ContainerStarted","Data":"530c95568bf4fb8d7a71cea521fe2192ce4046203842abc6cf7298b068f02b8a"} Nov 29 00:17:13 crc kubenswrapper[4931]: I1129 00:17:13.923222 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0d661705-7bb5-482f-b2a8-79b44f6351ac","Type":"ContainerStarted","Data":"4e5115ada1c17ffa4b9282df255656354f160bb6c4761063ca2718a5f5294d12"} Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.037322 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 00:17:14 crc kubenswrapper[4931]: E1129 00:17:14.196475 4931 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 29 00:17:14 crc kubenswrapper[4931]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/c3a0e1d0-d484-4b42-81a1-f0520ef32717/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 29 00:17:14 crc kubenswrapper[4931]: > podSandboxID="32b787ad030bda7449cdc19ba198cc2aff751d0a4ceae79815f83c386531d60d" Nov 29 00:17:14 crc kubenswrapper[4931]: E1129 00:17:14.196642 4931 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 29 00:17:14 crc kubenswrapper[4931]: init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jbbxx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-8mx5k_openstack(c3a0e1d0-d484-4b42-81a1-f0520ef32717): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/c3a0e1d0-d484-4b42-81a1-f0520ef32717/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 29 00:17:14 crc kubenswrapper[4931]: > logger="UnhandledError" Nov 29 00:17:14 crc kubenswrapper[4931]: E1129 00:17:14.197923 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/c3a0e1d0-d484-4b42-81a1-f0520ef32717/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" podUID="c3a0e1d0-d484-4b42-81a1-f0520ef32717" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.262879 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.315333 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.383488 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/866218de-4228-42aa-a27e-3da9e742a615-config\") pod \"866218de-4228-42aa-a27e-3da9e742a615\" (UID: \"866218de-4228-42aa-a27e-3da9e742a615\") " Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.384027 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xj29g\" (UniqueName: \"kubernetes.io/projected/866218de-4228-42aa-a27e-3da9e742a615-kube-api-access-xj29g\") pod \"866218de-4228-42aa-a27e-3da9e742a615\" (UID: \"866218de-4228-42aa-a27e-3da9e742a615\") " Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.384080 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2-config\") pod \"7cbb5cb5-d10a-4ee0-9345-fc5b651999d2\" (UID: \"7cbb5cb5-d10a-4ee0-9345-fc5b651999d2\") " Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.384125 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/866218de-4228-42aa-a27e-3da9e742a615-dns-svc\") pod \"866218de-4228-42aa-a27e-3da9e742a615\" (UID: \"866218de-4228-42aa-a27e-3da9e742a615\") " Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.384233 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x59sk\" (UniqueName: \"kubernetes.io/projected/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2-kube-api-access-x59sk\") pod \"7cbb5cb5-d10a-4ee0-9345-fc5b651999d2\" (UID: \"7cbb5cb5-d10a-4ee0-9345-fc5b651999d2\") " Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.384873 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2-config" (OuterVolumeSpecName: "config") pod "7cbb5cb5-d10a-4ee0-9345-fc5b651999d2" (UID: "7cbb5cb5-d10a-4ee0-9345-fc5b651999d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.384953 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/866218de-4228-42aa-a27e-3da9e742a615-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "866218de-4228-42aa-a27e-3da9e742a615" (UID: "866218de-4228-42aa-a27e-3da9e742a615"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.385095 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/866218de-4228-42aa-a27e-3da9e742a615-config" (OuterVolumeSpecName: "config") pod "866218de-4228-42aa-a27e-3da9e742a615" (UID: "866218de-4228-42aa-a27e-3da9e742a615"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.387866 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/866218de-4228-42aa-a27e-3da9e742a615-kube-api-access-xj29g" (OuterVolumeSpecName: "kube-api-access-xj29g") pod "866218de-4228-42aa-a27e-3da9e742a615" (UID: "866218de-4228-42aa-a27e-3da9e742a615"). InnerVolumeSpecName "kube-api-access-xj29g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.390135 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2-kube-api-access-x59sk" (OuterVolumeSpecName: "kube-api-access-x59sk") pod "7cbb5cb5-d10a-4ee0-9345-fc5b651999d2" (UID: "7cbb5cb5-d10a-4ee0-9345-fc5b651999d2"). InnerVolumeSpecName "kube-api-access-x59sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.486966 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x59sk\" (UniqueName: \"kubernetes.io/projected/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2-kube-api-access-x59sk\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.487065 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/866218de-4228-42aa-a27e-3da9e742a615-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.487081 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xj29g\" (UniqueName: \"kubernetes.io/projected/866218de-4228-42aa-a27e-3da9e742a615-kube-api-access-xj29g\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.487098 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.487108 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/866218de-4228-42aa-a27e-3da9e742a615-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.829844 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.932248 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"979903af-0c72-45fe-a3de-0dc52370bc57","Type":"ContainerStarted","Data":"c8b0964ad5542f30ff0a0ca983025dcbadc9e1c355931a7b28280d47f5643721"} Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.933711 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" event={"ID":"a0b54682-21aa-4382-b7c4-5db01da1cab8","Type":"ContainerStarted","Data":"813fce3f7ef18e8abaeaf642f1b9fbf0709577ebc61d60170d1c8e78c34790d7"} Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.934678 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.936454 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" event={"ID":"866218de-4228-42aa-a27e-3da9e742a615","Type":"ContainerDied","Data":"50740c3eeef2b323d0df957485fe4141eac9f4c0dfe2a97b36710cea91f9cb0c"} Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.936636 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-jdhdt" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.958215 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"be472357-f01e-4e83-8903-49c0e2f60c97","Type":"ContainerStarted","Data":"1361fab9af24f7ae14afba21dcb17eadfe875b9d0ecaaefdc540e541f7e313ef"} Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.958745 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" podStartSLOduration=2.692943176 podStartE2EDuration="18.958727206s" podCreationTimestamp="2025-11-29 00:16:56 +0000 UTC" firstStartedPulling="2025-11-29 00:16:56.912141607 +0000 UTC m=+1030.074034839" lastFinishedPulling="2025-11-29 00:17:13.177925637 +0000 UTC m=+1046.339818869" observedRunningTime="2025-11-29 00:17:14.952491083 +0000 UTC m=+1048.114384335" watchObservedRunningTime="2025-11-29 00:17:14.958727206 +0000 UTC m=+1048.120620458" Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.959288 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" event={"ID":"7cbb5cb5-d10a-4ee0-9345-fc5b651999d2","Type":"ContainerDied","Data":"06611ae94cbd1e0ed38277714441981dd3c3129a35aa10d5027019cef5b8cb24"} Nov 29 00:17:14 crc kubenswrapper[4931]: I1129 00:17:14.959315 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tghf9" Nov 29 00:17:15 crc kubenswrapper[4931]: I1129 00:17:15.027695 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-jdhdt"] Nov 29 00:17:15 crc kubenswrapper[4931]: I1129 00:17:15.034140 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-jdhdt"] Nov 29 00:17:15 crc kubenswrapper[4931]: I1129 00:17:15.071883 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tghf9"] Nov 29 00:17:15 crc kubenswrapper[4931]: I1129 00:17:15.076210 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tghf9"] Nov 29 00:17:15 crc kubenswrapper[4931]: I1129 00:17:15.226058 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cbb5cb5-d10a-4ee0-9345-fc5b651999d2" path="/var/lib/kubelet/pods/7cbb5cb5-d10a-4ee0-9345-fc5b651999d2/volumes" Nov 29 00:17:15 crc kubenswrapper[4931]: I1129 00:17:15.226526 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="866218de-4228-42aa-a27e-3da9e742a615" path="/var/lib/kubelet/pods/866218de-4228-42aa-a27e-3da9e742a615/volumes" Nov 29 00:17:21 crc kubenswrapper[4931]: I1129 00:17:21.494012 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:17:21 crc kubenswrapper[4931]: I1129 00:17:21.549762 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-8mx5k"] Nov 29 00:17:23 crc kubenswrapper[4931]: I1129 00:17:23.735318 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:17:23 crc kubenswrapper[4931]: I1129 00:17:23.735986 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:17:23 crc kubenswrapper[4931]: I1129 00:17:23.736044 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:17:23 crc kubenswrapper[4931]: I1129 00:17:23.737006 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"63e24d3607eb4b6f4826de9209c8d28131ebf746bafba6dd38a837ce70168944"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 00:17:23 crc kubenswrapper[4931]: I1129 00:17:23.737090 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://63e24d3607eb4b6f4826de9209c8d28131ebf746bafba6dd38a837ce70168944" gracePeriod=600 Nov 29 00:17:23 crc kubenswrapper[4931]: I1129 00:17:23.963608 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.038977 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="63e24d3607eb4b6f4826de9209c8d28131ebf746bafba6dd38a837ce70168944" exitCode=0 Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.039054 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"63e24d3607eb4b6f4826de9209c8d28131ebf746bafba6dd38a837ce70168944"} Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.039105 4931 scope.go:117] "RemoveContainer" containerID="b74c2da42faa59386db14c3bb51aa4674c697dd1c4383118b269448576e97f17" Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.040913 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" event={"ID":"c3a0e1d0-d484-4b42-81a1-f0520ef32717","Type":"ContainerDied","Data":"32b787ad030bda7449cdc19ba198cc2aff751d0a4ceae79815f83c386531d60d"} Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.040988 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-8mx5k" Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.057769 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbbxx\" (UniqueName: \"kubernetes.io/projected/c3a0e1d0-d484-4b42-81a1-f0520ef32717-kube-api-access-jbbxx\") pod \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\" (UID: \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\") " Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.057884 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3a0e1d0-d484-4b42-81a1-f0520ef32717-dns-svc\") pod \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\" (UID: \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\") " Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.058016 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a0e1d0-d484-4b42-81a1-f0520ef32717-config\") pod \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\" (UID: \"c3a0e1d0-d484-4b42-81a1-f0520ef32717\") " Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.062982 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3a0e1d0-d484-4b42-81a1-f0520ef32717-kube-api-access-jbbxx" (OuterVolumeSpecName: "kube-api-access-jbbxx") pod "c3a0e1d0-d484-4b42-81a1-f0520ef32717" (UID: "c3a0e1d0-d484-4b42-81a1-f0520ef32717"). InnerVolumeSpecName "kube-api-access-jbbxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.075772 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3a0e1d0-d484-4b42-81a1-f0520ef32717-config" (OuterVolumeSpecName: "config") pod "c3a0e1d0-d484-4b42-81a1-f0520ef32717" (UID: "c3a0e1d0-d484-4b42-81a1-f0520ef32717"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.075960 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3a0e1d0-d484-4b42-81a1-f0520ef32717-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c3a0e1d0-d484-4b42-81a1-f0520ef32717" (UID: "c3a0e1d0-d484-4b42-81a1-f0520ef32717"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.160676 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3a0e1d0-d484-4b42-81a1-f0520ef32717-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.160724 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3a0e1d0-d484-4b42-81a1-f0520ef32717-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.160745 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbbxx\" (UniqueName: \"kubernetes.io/projected/c3a0e1d0-d484-4b42-81a1-f0520ef32717-kube-api-access-jbbxx\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.394653 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-8mx5k"] Nov 29 00:17:24 crc kubenswrapper[4931]: I1129 00:17:24.401353 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-8mx5k"] Nov 29 00:17:25 crc kubenswrapper[4931]: I1129 00:17:25.237599 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3a0e1d0-d484-4b42-81a1-f0520ef32717" path="/var/lib/kubelet/pods/c3a0e1d0-d484-4b42-81a1-f0520ef32717/volumes" Nov 29 00:17:27 crc kubenswrapper[4931]: I1129 00:17:27.077274 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"7ca6926ea091667ee0b6b53772fafdc3830c2a5a094ad559af9957d016e82d86"} Nov 29 00:17:29 crc kubenswrapper[4931]: I1129 00:17:29.093157 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"37afe9f6-664f-4f51-90a9-485e87a1db92","Type":"ContainerStarted","Data":"f4f4a8788742a1e767820afb61aa0711131dacad8ae576939e94f248328825e5"} Nov 29 00:17:29 crc kubenswrapper[4931]: I1129 00:17:29.095726 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ba2777f2-23db-46b8-8ccf-cb0813a39678","Type":"ContainerStarted","Data":"171e32b8938ce416899eb7164e915aff9a40c8f51a6a2e05fc77206d5afbb0ec"} Nov 29 00:17:29 crc kubenswrapper[4931]: I1129 00:17:29.097407 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"979903af-0c72-45fe-a3de-0dc52370bc57","Type":"ContainerStarted","Data":"1c4f0337274dc7bc55dcd9ea12de2271569763d813a9c19963c459b4a194ad6e"} Nov 29 00:17:29 crc kubenswrapper[4931]: I1129 00:17:29.100585 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7s9fx" event={"ID":"28c11b0b-3026-4f9f-8de8-71174fbd19d6","Type":"ContainerStarted","Data":"457635f6e39c579afae6b078e0686c7cff8781944144f7821001d49e3d87c27d"} Nov 29 00:17:29 crc kubenswrapper[4931]: I1129 00:17:29.102002 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d60ec023-b0d6-466c-b188-5e0e2b47a962","Type":"ContainerStarted","Data":"6fa1c435a4814003963366602bea181fed4eca49e7ef0218d17019797f58c035"} Nov 29 00:17:29 crc kubenswrapper[4931]: I1129 00:17:29.102190 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 29 00:17:29 crc kubenswrapper[4931]: I1129 00:17:29.178961 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=17.308228253 podStartE2EDuration="29.178943256s" podCreationTimestamp="2025-11-29 00:17:00 +0000 UTC" firstStartedPulling="2025-11-29 00:17:13.829758327 +0000 UTC m=+1046.991651559" lastFinishedPulling="2025-11-29 00:17:25.70047333 +0000 UTC m=+1058.862366562" observedRunningTime="2025-11-29 00:17:29.174554594 +0000 UTC m=+1062.336447866" watchObservedRunningTime="2025-11-29 00:17:29.178943256 +0000 UTC m=+1062.340836488" Nov 29 00:17:30 crc kubenswrapper[4931]: I1129 00:17:30.116439 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c2b4d256-9b86-48e3-98d8-266271a338dd","Type":"ContainerStarted","Data":"f93d3096779e6ed08a2f27aad5086859f237ac05e9f71a441679105de64b0d03"} Nov 29 00:17:30 crc kubenswrapper[4931]: I1129 00:17:30.119054 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fbd78afc-2d6d-4118-92a2-e3499560359f","Type":"ContainerStarted","Data":"7fe4cd6379e10e7b3ca34aac9701c07da875d4602076ce55cb8db09358f686be"} Nov 29 00:17:30 crc kubenswrapper[4931]: I1129 00:17:30.122249 4931 generic.go:334] "Generic (PLEG): container finished" podID="28c11b0b-3026-4f9f-8de8-71174fbd19d6" containerID="457635f6e39c579afae6b078e0686c7cff8781944144f7821001d49e3d87c27d" exitCode=0 Nov 29 00:17:30 crc kubenswrapper[4931]: I1129 00:17:30.122327 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7s9fx" event={"ID":"28c11b0b-3026-4f9f-8de8-71174fbd19d6","Type":"ContainerDied","Data":"457635f6e39c579afae6b078e0686c7cff8781944144f7821001d49e3d87c27d"} Nov 29 00:17:30 crc kubenswrapper[4931]: I1129 00:17:30.123951 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"be472357-f01e-4e83-8903-49c0e2f60c97","Type":"ContainerStarted","Data":"456cc2174afa59dc77de92827cdb152c0ca1ed767f352bf123447d6dd0c91c71"} Nov 29 00:17:30 crc kubenswrapper[4931]: I1129 00:17:30.125249 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 29 00:17:30 crc kubenswrapper[4931]: I1129 00:17:30.128257 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-trsph" event={"ID":"a64bc6bd-8033-45c9-b2d7-9602aea09225","Type":"ContainerStarted","Data":"f586203e49ac4b939fa80a34c9ec98176f9f95ce93f51474910d40caa2748f6c"} Nov 29 00:17:30 crc kubenswrapper[4931]: I1129 00:17:30.128664 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-trsph" Nov 29 00:17:30 crc kubenswrapper[4931]: I1129 00:17:30.131646 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0d661705-7bb5-482f-b2a8-79b44f6351ac","Type":"ContainerStarted","Data":"c13dfad395ef5b782b6371d50ee00ee1d0b2fdc386adbf7dfeb5c206bbf2019d"} Nov 29 00:17:30 crc kubenswrapper[4931]: I1129 00:17:30.186328 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-trsph" podStartSLOduration=12.028279435 podStartE2EDuration="24.186312118s" podCreationTimestamp="2025-11-29 00:17:06 +0000 UTC" firstStartedPulling="2025-11-29 00:17:13.820661935 +0000 UTC m=+1046.982555167" lastFinishedPulling="2025-11-29 00:17:25.978694588 +0000 UTC m=+1059.140587850" observedRunningTime="2025-11-29 00:17:30.181549306 +0000 UTC m=+1063.343442538" watchObservedRunningTime="2025-11-29 00:17:30.186312118 +0000 UTC m=+1063.348205350" Nov 29 00:17:30 crc kubenswrapper[4931]: I1129 00:17:30.266208 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=13.817763728 podStartE2EDuration="28.266190501s" podCreationTimestamp="2025-11-29 00:17:02 +0000 UTC" firstStartedPulling="2025-11-29 00:17:14.056427617 +0000 UTC m=+1047.218320849" lastFinishedPulling="2025-11-29 00:17:28.50485439 +0000 UTC m=+1061.666747622" observedRunningTime="2025-11-29 00:17:30.265403179 +0000 UTC m=+1063.427296411" watchObservedRunningTime="2025-11-29 00:17:30.266190501 +0000 UTC m=+1063.428083733" Nov 29 00:17:31 crc kubenswrapper[4931]: I1129 00:17:31.146039 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7s9fx" event={"ID":"28c11b0b-3026-4f9f-8de8-71174fbd19d6","Type":"ContainerStarted","Data":"d3e528507c40d89f2f4ae0a99d635e163287d23b3a351cd708c4536da45e9ee6"} Nov 29 00:17:31 crc kubenswrapper[4931]: I1129 00:17:31.146480 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7s9fx" event={"ID":"28c11b0b-3026-4f9f-8de8-71174fbd19d6","Type":"ContainerStarted","Data":"72c0b25aeebf8aeae1986ff1bba1b7fb573b2c3e282ffbf748ed31510bb1102b"} Nov 29 00:17:31 crc kubenswrapper[4931]: I1129 00:17:31.844706 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:31 crc kubenswrapper[4931]: I1129 00:17:31.844999 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:17:33 crc kubenswrapper[4931]: I1129 00:17:33.170630 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"37afe9f6-664f-4f51-90a9-485e87a1db92","Type":"ContainerStarted","Data":"77d48e27acbe941678c99f64833b888e6b3c88bb42428c3c6fb01d6f4e466008"} Nov 29 00:17:33 crc kubenswrapper[4931]: I1129 00:17:33.172562 4931 generic.go:334] "Generic (PLEG): container finished" podID="ba2777f2-23db-46b8-8ccf-cb0813a39678" containerID="171e32b8938ce416899eb7164e915aff9a40c8f51a6a2e05fc77206d5afbb0ec" exitCode=0 Nov 29 00:17:33 crc kubenswrapper[4931]: I1129 00:17:33.172638 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ba2777f2-23db-46b8-8ccf-cb0813a39678","Type":"ContainerDied","Data":"171e32b8938ce416899eb7164e915aff9a40c8f51a6a2e05fc77206d5afbb0ec"} Nov 29 00:17:33 crc kubenswrapper[4931]: I1129 00:17:33.176353 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"979903af-0c72-45fe-a3de-0dc52370bc57","Type":"ContainerStarted","Data":"21043e6d97777d49a1fc8eadd78c1bd8b92b81414e28eb057a755b17eeb0efae"} Nov 29 00:17:33 crc kubenswrapper[4931]: I1129 00:17:33.178732 4931 generic.go:334] "Generic (PLEG): container finished" podID="c2b4d256-9b86-48e3-98d8-266271a338dd" containerID="f93d3096779e6ed08a2f27aad5086859f237ac05e9f71a441679105de64b0d03" exitCode=0 Nov 29 00:17:33 crc kubenswrapper[4931]: I1129 00:17:33.179451 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c2b4d256-9b86-48e3-98d8-266271a338dd","Type":"ContainerDied","Data":"f93d3096779e6ed08a2f27aad5086859f237ac05e9f71a441679105de64b0d03"} Nov 29 00:17:33 crc kubenswrapper[4931]: I1129 00:17:33.205290 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=7.038596138 podStartE2EDuration="25.205273852s" podCreationTimestamp="2025-11-29 00:17:08 +0000 UTC" firstStartedPulling="2025-11-29 00:17:13.897115303 +0000 UTC m=+1047.059008535" lastFinishedPulling="2025-11-29 00:17:32.063793017 +0000 UTC m=+1065.225686249" observedRunningTime="2025-11-29 00:17:33.200390587 +0000 UTC m=+1066.362283839" watchObservedRunningTime="2025-11-29 00:17:33.205273852 +0000 UTC m=+1066.367167074" Nov 29 00:17:33 crc kubenswrapper[4931]: I1129 00:17:33.205996 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-7s9fx" podStartSLOduration=15.793929356 podStartE2EDuration="27.205988392s" podCreationTimestamp="2025-11-29 00:17:06 +0000 UTC" firstStartedPulling="2025-11-29 00:17:13.817058455 +0000 UTC m=+1046.978951677" lastFinishedPulling="2025-11-29 00:17:25.229117481 +0000 UTC m=+1058.391010713" observedRunningTime="2025-11-29 00:17:31.170297901 +0000 UTC m=+1064.332191123" watchObservedRunningTime="2025-11-29 00:17:33.205988392 +0000 UTC m=+1066.367881624" Nov 29 00:17:33 crc kubenswrapper[4931]: I1129 00:17:33.254888 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=11.015612448 podStartE2EDuration="28.254868427s" podCreationTimestamp="2025-11-29 00:17:05 +0000 UTC" firstStartedPulling="2025-11-29 00:17:14.83977966 +0000 UTC m=+1048.001672892" lastFinishedPulling="2025-11-29 00:17:32.079035639 +0000 UTC m=+1065.240928871" observedRunningTime="2025-11-29 00:17:33.240976652 +0000 UTC m=+1066.402869894" watchObservedRunningTime="2025-11-29 00:17:33.254868427 +0000 UTC m=+1066.416761669" Nov 29 00:17:33 crc kubenswrapper[4931]: I1129 00:17:33.898539 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:33 crc kubenswrapper[4931]: I1129 00:17:33.969310 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.192793 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ba2777f2-23db-46b8-8ccf-cb0813a39678","Type":"ContainerStarted","Data":"9a9eab2f2eeecf14f98758745f0de3cffd68e906fd6ca9de2e049b194d9637e9"} Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.196252 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c2b4d256-9b86-48e3-98d8-266271a338dd","Type":"ContainerStarted","Data":"78b1c5e3ad127c09847a64f4e0e80f5f6effa7a54f48bf2489e7619024110259"} Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.196932 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.221246 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=24.913176111 podStartE2EDuration="37.221220291s" podCreationTimestamp="2025-11-29 00:16:57 +0000 UTC" firstStartedPulling="2025-11-29 00:17:13.668848478 +0000 UTC m=+1046.830741730" lastFinishedPulling="2025-11-29 00:17:25.976892678 +0000 UTC m=+1059.138785910" observedRunningTime="2025-11-29 00:17:34.218315641 +0000 UTC m=+1067.380208883" watchObservedRunningTime="2025-11-29 00:17:34.221220291 +0000 UTC m=+1067.383113583" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.247030 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=23.693734067 podStartE2EDuration="35.247012166s" podCreationTimestamp="2025-11-29 00:16:59 +0000 UTC" firstStartedPulling="2025-11-29 00:17:13.675111332 +0000 UTC m=+1046.837004574" lastFinishedPulling="2025-11-29 00:17:25.228389401 +0000 UTC m=+1058.390282673" observedRunningTime="2025-11-29 00:17:34.24317722 +0000 UTC m=+1067.405070462" watchObservedRunningTime="2025-11-29 00:17:34.247012166 +0000 UTC m=+1067.408905398" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.249097 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.263025 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.295219 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.522369 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-sbpv2"] Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.523882 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.528082 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.544486 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-sbpv2"] Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.671535 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-sbpv2\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.671586 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-config\") pod \"dnsmasq-dns-6bc7876d45-sbpv2\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.671875 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-sbpv2\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.671945 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrfzh\" (UniqueName: \"kubernetes.io/projected/2111b30e-5f94-41e6-9ff6-dbc1101b2799-kube-api-access-jrfzh\") pod \"dnsmasq-dns-6bc7876d45-sbpv2\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.686996 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-m9dkp"] Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.687912 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.690972 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.704397 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-m9dkp"] Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.773323 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-sbpv2\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.773368 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-config\") pod \"dnsmasq-dns-6bc7876d45-sbpv2\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.773392 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f9b49437-b579-4f61-82a3-57d81e77b0b7-ovs-rundir\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.773413 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvktb\" (UniqueName: \"kubernetes.io/projected/f9b49437-b579-4f61-82a3-57d81e77b0b7-kube-api-access-pvktb\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.773444 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f9b49437-b579-4f61-82a3-57d81e77b0b7-ovn-rundir\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.773639 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9b49437-b579-4f61-82a3-57d81e77b0b7-combined-ca-bundle\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.773721 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-sbpv2\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.773758 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrfzh\" (UniqueName: \"kubernetes.io/projected/2111b30e-5f94-41e6-9ff6-dbc1101b2799-kube-api-access-jrfzh\") pod \"dnsmasq-dns-6bc7876d45-sbpv2\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.773794 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9b49437-b579-4f61-82a3-57d81e77b0b7-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.773836 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9b49437-b579-4f61-82a3-57d81e77b0b7-config\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.774167 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-config\") pod \"dnsmasq-dns-6bc7876d45-sbpv2\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.774950 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-sbpv2\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.775201 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-sbpv2\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.791601 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrfzh\" (UniqueName: \"kubernetes.io/projected/2111b30e-5f94-41e6-9ff6-dbc1101b2799-kube-api-access-jrfzh\") pod \"dnsmasq-dns-6bc7876d45-sbpv2\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.845676 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.877577 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f9b49437-b579-4f61-82a3-57d81e77b0b7-ovn-rundir\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.877966 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9b49437-b579-4f61-82a3-57d81e77b0b7-combined-ca-bundle\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.878018 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9b49437-b579-4f61-82a3-57d81e77b0b7-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.878039 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9b49437-b579-4f61-82a3-57d81e77b0b7-config\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.878093 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f9b49437-b579-4f61-82a3-57d81e77b0b7-ovs-rundir\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.878115 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvktb\" (UniqueName: \"kubernetes.io/projected/f9b49437-b579-4f61-82a3-57d81e77b0b7-kube-api-access-pvktb\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.878629 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f9b49437-b579-4f61-82a3-57d81e77b0b7-ovn-rundir\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.879234 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f9b49437-b579-4f61-82a3-57d81e77b0b7-ovs-rundir\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.879762 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9b49437-b579-4f61-82a3-57d81e77b0b7-config\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.884355 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9b49437-b579-4f61-82a3-57d81e77b0b7-combined-ca-bundle\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.884726 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9b49437-b579-4f61-82a3-57d81e77b0b7-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.897573 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvktb\" (UniqueName: \"kubernetes.io/projected/f9b49437-b579-4f61-82a3-57d81e77b0b7-kube-api-access-pvktb\") pod \"ovn-controller-metrics-m9dkp\" (UID: \"f9b49437-b579-4f61-82a3-57d81e77b0b7\") " pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.924249 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-sbpv2"] Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.957047 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-cq95h"] Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.958548 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.962248 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 29 00:17:34 crc kubenswrapper[4931]: I1129 00:17:34.964447 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-cq95h"] Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.002669 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-m9dkp" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.080697 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-config\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.081299 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.081348 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc7fr\" (UniqueName: \"kubernetes.io/projected/1b45e505-3940-45d0-868e-604f8490f8d2-kube-api-access-lc7fr\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.081392 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-dns-svc\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.081422 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.183177 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-config\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.183241 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.183299 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc7fr\" (UniqueName: \"kubernetes.io/projected/1b45e505-3940-45d0-868e-604f8490f8d2-kube-api-access-lc7fr\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.183391 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-dns-svc\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.183858 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.184300 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-dns-svc\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.184550 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-config\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.184738 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.185432 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.201378 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc7fr\" (UniqueName: \"kubernetes.io/projected/1b45e505-3940-45d0-868e-604f8490f8d2-kube-api-access-lc7fr\") pod \"dnsmasq-dns-8554648995-cq95h\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.209249 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.259580 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.314206 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.356238 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-sbpv2"] Nov 29 00:17:35 crc kubenswrapper[4931]: W1129 00:17:35.360566 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2111b30e_5f94_41e6_9ff6_dbc1101b2799.slice/crio-0d7d1fc0ac359b8aaa86830ad818999f080755086b5c26107951c552262754ba WatchSource:0}: Error finding container 0d7d1fc0ac359b8aaa86830ad818999f080755086b5c26107951c552262754ba: Status 404 returned error can't find the container with id 0d7d1fc0ac359b8aaa86830ad818999f080755086b5c26107951c552262754ba Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.446834 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-m9dkp"] Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.511501 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.529378 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.540444 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-kqgfl" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.540925 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.542140 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.545144 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.587473 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.591869 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.591973 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-config\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.592008 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-scripts\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.592150 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.592272 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.592305 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vz97\" (UniqueName: \"kubernetes.io/projected/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-kube-api-access-4vz97\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.592406 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.696055 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.696124 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.696141 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vz97\" (UniqueName: \"kubernetes.io/projected/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-kube-api-access-4vz97\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.696172 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.696205 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.696230 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-config\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.696252 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-scripts\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.698000 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-config\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.698741 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.698990 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-scripts\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.703325 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.703613 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.709694 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.716589 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vz97\" (UniqueName: \"kubernetes.io/projected/d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb-kube-api-access-4vz97\") pod \"ovn-northd-0\" (UID: \"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb\") " pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.805569 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-cq95h"] Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.863990 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 29 00:17:35 crc kubenswrapper[4931]: I1129 00:17:35.959996 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.217765 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-m9dkp" event={"ID":"f9b49437-b579-4f61-82a3-57d81e77b0b7","Type":"ContainerStarted","Data":"03c421cb85f815cf2c33503e3e49dde5b3eba6b1c0d6ce7ca35dffffb32ecaf6"} Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.218036 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-m9dkp" event={"ID":"f9b49437-b579-4f61-82a3-57d81e77b0b7","Type":"ContainerStarted","Data":"a9aef61264a73298666b7ba878c9854f28007c3357535cd93490a43b4999ce33"} Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.219464 4931 generic.go:334] "Generic (PLEG): container finished" podID="1b45e505-3940-45d0-868e-604f8490f8d2" containerID="83f041170c695791d74550dbca6f2c17a1328731a80329064eef3276919fd45e" exitCode=0 Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.219509 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-cq95h" event={"ID":"1b45e505-3940-45d0-868e-604f8490f8d2","Type":"ContainerDied","Data":"83f041170c695791d74550dbca6f2c17a1328731a80329064eef3276919fd45e"} Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.219524 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-cq95h" event={"ID":"1b45e505-3940-45d0-868e-604f8490f8d2","Type":"ContainerStarted","Data":"3a2167d0e06b36233a70da9bc3aa6f18824e70ab174ff5c4af99257213480794"} Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.222561 4931 generic.go:334] "Generic (PLEG): container finished" podID="2111b30e-5f94-41e6-9ff6-dbc1101b2799" containerID="a433265c45c770089e37bafd46afd74c0d5167cd994d9798c3df71a3fbd39d02" exitCode=0 Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.223590 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" event={"ID":"2111b30e-5f94-41e6-9ff6-dbc1101b2799","Type":"ContainerDied","Data":"a433265c45c770089e37bafd46afd74c0d5167cd994d9798c3df71a3fbd39d02"} Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.223631 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" event={"ID":"2111b30e-5f94-41e6-9ff6-dbc1101b2799","Type":"ContainerStarted","Data":"0d7d1fc0ac359b8aaa86830ad818999f080755086b5c26107951c552262754ba"} Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.252367 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-m9dkp" podStartSLOduration=2.252345987 podStartE2EDuration="2.252345987s" podCreationTimestamp="2025-11-29 00:17:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:17:36.247797751 +0000 UTC m=+1069.409690983" watchObservedRunningTime="2025-11-29 00:17:36.252345987 +0000 UTC m=+1069.414239219" Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.372740 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 29 00:17:36 crc kubenswrapper[4931]: W1129 00:17:36.379564 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7c4acbb_4e74_4467_a524_bc9cf9bb4fdb.slice/crio-5153bb5e992acf09af1a93e6a24140f4c784b2ee60ea94c61c71fb04ddc4de31 WatchSource:0}: Error finding container 5153bb5e992acf09af1a93e6a24140f4c784b2ee60ea94c61c71fb04ddc4de31: Status 404 returned error can't find the container with id 5153bb5e992acf09af1a93e6a24140f4c784b2ee60ea94c61c71fb04ddc4de31 Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.475159 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.519212 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-ovsdbserver-sb\") pod \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.519307 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-dns-svc\") pod \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.519470 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrfzh\" (UniqueName: \"kubernetes.io/projected/2111b30e-5f94-41e6-9ff6-dbc1101b2799-kube-api-access-jrfzh\") pod \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.519512 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-config\") pod \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\" (UID: \"2111b30e-5f94-41e6-9ff6-dbc1101b2799\") " Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.525344 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2111b30e-5f94-41e6-9ff6-dbc1101b2799-kube-api-access-jrfzh" (OuterVolumeSpecName: "kube-api-access-jrfzh") pod "2111b30e-5f94-41e6-9ff6-dbc1101b2799" (UID: "2111b30e-5f94-41e6-9ff6-dbc1101b2799"). InnerVolumeSpecName "kube-api-access-jrfzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.536637 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-config" (OuterVolumeSpecName: "config") pod "2111b30e-5f94-41e6-9ff6-dbc1101b2799" (UID: "2111b30e-5f94-41e6-9ff6-dbc1101b2799"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.543358 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2111b30e-5f94-41e6-9ff6-dbc1101b2799" (UID: "2111b30e-5f94-41e6-9ff6-dbc1101b2799"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.545277 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2111b30e-5f94-41e6-9ff6-dbc1101b2799" (UID: "2111b30e-5f94-41e6-9ff6-dbc1101b2799"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.620719 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.620764 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrfzh\" (UniqueName: \"kubernetes.io/projected/2111b30e-5f94-41e6-9ff6-dbc1101b2799-kube-api-access-jrfzh\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.620776 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:36 crc kubenswrapper[4931]: I1129 00:17:36.620789 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2111b30e-5f94-41e6-9ff6-dbc1101b2799-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:37 crc kubenswrapper[4931]: I1129 00:17:37.276032 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb","Type":"ContainerStarted","Data":"5153bb5e992acf09af1a93e6a24140f4c784b2ee60ea94c61c71fb04ddc4de31"} Nov 29 00:17:37 crc kubenswrapper[4931]: I1129 00:17:37.303547 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-cq95h" event={"ID":"1b45e505-3940-45d0-868e-604f8490f8d2","Type":"ContainerStarted","Data":"61e39270fc93f60ec85d193da8dd4a5da50fa24061071ba03da712dcd31bd678"} Nov 29 00:17:37 crc kubenswrapper[4931]: I1129 00:17:37.304483 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:37 crc kubenswrapper[4931]: I1129 00:17:37.308711 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" Nov 29 00:17:37 crc kubenswrapper[4931]: I1129 00:17:37.308934 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-sbpv2" event={"ID":"2111b30e-5f94-41e6-9ff6-dbc1101b2799","Type":"ContainerDied","Data":"0d7d1fc0ac359b8aaa86830ad818999f080755086b5c26107951c552262754ba"} Nov 29 00:17:37 crc kubenswrapper[4931]: I1129 00:17:37.308982 4931 scope.go:117] "RemoveContainer" containerID="a433265c45c770089e37bafd46afd74c0d5167cd994d9798c3df71a3fbd39d02" Nov 29 00:17:37 crc kubenswrapper[4931]: I1129 00:17:37.328390 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-cq95h" podStartSLOduration=3.328369159 podStartE2EDuration="3.328369159s" podCreationTimestamp="2025-11-29 00:17:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:17:37.32480209 +0000 UTC m=+1070.486695322" watchObservedRunningTime="2025-11-29 00:17:37.328369159 +0000 UTC m=+1070.490262391" Nov 29 00:17:37 crc kubenswrapper[4931]: I1129 00:17:37.378952 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-sbpv2"] Nov 29 00:17:37 crc kubenswrapper[4931]: I1129 00:17:37.391614 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-sbpv2"] Nov 29 00:17:38 crc kubenswrapper[4931]: I1129 00:17:38.321509 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb","Type":"ContainerStarted","Data":"cee80168bc9246ca6b3261255437d3090a26968f611e5a5ddf8ca991e7626729"} Nov 29 00:17:38 crc kubenswrapper[4931]: I1129 00:17:38.322285 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb","Type":"ContainerStarted","Data":"05b48b97232c2134e626731f1e72b4aebb7d1b54e6d77ce292dbf5f4a619d40b"} Nov 29 00:17:38 crc kubenswrapper[4931]: I1129 00:17:38.358531 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.1378032989999998 podStartE2EDuration="3.35850422s" podCreationTimestamp="2025-11-29 00:17:35 +0000 UTC" firstStartedPulling="2025-11-29 00:17:36.386544774 +0000 UTC m=+1069.548438006" lastFinishedPulling="2025-11-29 00:17:37.607245695 +0000 UTC m=+1070.769138927" observedRunningTime="2025-11-29 00:17:38.351736813 +0000 UTC m=+1071.513630055" watchObservedRunningTime="2025-11-29 00:17:38.35850422 +0000 UTC m=+1071.520397492" Nov 29 00:17:39 crc kubenswrapper[4931]: I1129 00:17:39.036847 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 29 00:17:39 crc kubenswrapper[4931]: I1129 00:17:39.036920 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 29 00:17:39 crc kubenswrapper[4931]: I1129 00:17:39.222352 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2111b30e-5f94-41e6-9ff6-dbc1101b2799" path="/var/lib/kubelet/pods/2111b30e-5f94-41e6-9ff6-dbc1101b2799/volumes" Nov 29 00:17:39 crc kubenswrapper[4931]: I1129 00:17:39.329062 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 29 00:17:40 crc kubenswrapper[4931]: I1129 00:17:40.580671 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:40 crc kubenswrapper[4931]: I1129 00:17:40.581951 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:40 crc kubenswrapper[4931]: I1129 00:17:40.691395 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:41 crc kubenswrapper[4931]: I1129 00:17:41.246534 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 29 00:17:41 crc kubenswrapper[4931]: I1129 00:17:41.326261 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 29 00:17:41 crc kubenswrapper[4931]: I1129 00:17:41.444776 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.616186 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.713082 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-cq95h"] Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.713575 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-cq95h" podUID="1b45e505-3940-45d0-868e-604f8490f8d2" containerName="dnsmasq-dns" containerID="cri-o://61e39270fc93f60ec85d193da8dd4a5da50fa24061071ba03da712dcd31bd678" gracePeriod=10 Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.714980 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.746970 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-z4dhc"] Nov 29 00:17:42 crc kubenswrapper[4931]: E1129 00:17:42.747278 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2111b30e-5f94-41e6-9ff6-dbc1101b2799" containerName="init" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.747295 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2111b30e-5f94-41e6-9ff6-dbc1101b2799" containerName="init" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.747450 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2111b30e-5f94-41e6-9ff6-dbc1101b2799" containerName="init" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.753091 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.790972 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-z4dhc"] Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.833548 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq2k4\" (UniqueName: \"kubernetes.io/projected/1f406c3b-dbb1-421e-9cee-334414d6089f-kube-api-access-zq2k4\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.833638 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-config\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.833669 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.833731 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.833774 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.935872 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-config\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.935922 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.935993 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.936042 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.936219 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq2k4\" (UniqueName: \"kubernetes.io/projected/1f406c3b-dbb1-421e-9cee-334414d6089f-kube-api-access-zq2k4\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.936985 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.937051 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-config\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.937066 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.936983 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:42 crc kubenswrapper[4931]: I1129 00:17:42.955525 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq2k4\" (UniqueName: \"kubernetes.io/projected/1f406c3b-dbb1-421e-9cee-334414d6089f-kube-api-access-zq2k4\") pod \"dnsmasq-dns-b8fbc5445-z4dhc\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.082367 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.362178 4931 generic.go:334] "Generic (PLEG): container finished" podID="1b45e505-3940-45d0-868e-604f8490f8d2" containerID="61e39270fc93f60ec85d193da8dd4a5da50fa24061071ba03da712dcd31bd678" exitCode=0 Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.362563 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-cq95h" event={"ID":"1b45e505-3940-45d0-868e-604f8490f8d2","Type":"ContainerDied","Data":"61e39270fc93f60ec85d193da8dd4a5da50fa24061071ba03da712dcd31bd678"} Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.568792 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-z4dhc"] Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.676234 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.816309 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 29 00:17:43 crc kubenswrapper[4931]: E1129 00:17:43.816937 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b45e505-3940-45d0-868e-604f8490f8d2" containerName="dnsmasq-dns" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.816959 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b45e505-3940-45d0-868e-604f8490f8d2" containerName="dnsmasq-dns" Nov 29 00:17:43 crc kubenswrapper[4931]: E1129 00:17:43.816984 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b45e505-3940-45d0-868e-604f8490f8d2" containerName="init" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.816993 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b45e505-3940-45d0-868e-604f8490f8d2" containerName="init" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.817178 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b45e505-3940-45d0-868e-604f8490f8d2" containerName="dnsmasq-dns" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.824759 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.827595 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.827912 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.836998 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-mmzf8" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.837116 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.857849 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-dns-svc\") pod \"1b45e505-3940-45d0-868e-604f8490f8d2\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.858235 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-config\") pod \"1b45e505-3940-45d0-868e-604f8490f8d2\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.858348 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-ovsdbserver-nb\") pod \"1b45e505-3940-45d0-868e-604f8490f8d2\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.858474 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-ovsdbserver-sb\") pod \"1b45e505-3940-45d0-868e-604f8490f8d2\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.858591 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc7fr\" (UniqueName: \"kubernetes.io/projected/1b45e505-3940-45d0-868e-604f8490f8d2-kube-api-access-lc7fr\") pod \"1b45e505-3940-45d0-868e-604f8490f8d2\" (UID: \"1b45e505-3940-45d0-868e-604f8490f8d2\") " Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.874800 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.880986 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b45e505-3940-45d0-868e-604f8490f8d2-kube-api-access-lc7fr" (OuterVolumeSpecName: "kube-api-access-lc7fr") pod "1b45e505-3940-45d0-868e-604f8490f8d2" (UID: "1b45e505-3940-45d0-868e-604f8490f8d2"). InnerVolumeSpecName "kube-api-access-lc7fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.908378 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1b45e505-3940-45d0-868e-604f8490f8d2" (UID: "1b45e505-3940-45d0-868e-604f8490f8d2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.911779 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1b45e505-3940-45d0-868e-604f8490f8d2" (UID: "1b45e505-3940-45d0-868e-604f8490f8d2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.916367 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-config" (OuterVolumeSpecName: "config") pod "1b45e505-3940-45d0-868e-604f8490f8d2" (UID: "1b45e505-3940-45d0-868e-604f8490f8d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.921600 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b45e505-3940-45d0-868e-604f8490f8d2" (UID: "1b45e505-3940-45d0-868e-604f8490f8d2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.960752 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.960837 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a720d879-8b09-452d-8371-9bf75f28cbde-lock\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.960892 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.960923 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntrtf\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-kube-api-access-ntrtf\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.960953 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a720d879-8b09-452d-8371-9bf75f28cbde-cache\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.961024 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.961034 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.961043 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.961052 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b45e505-3940-45d0-868e-604f8490f8d2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:43 crc kubenswrapper[4931]: I1129 00:17:43.961063 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc7fr\" (UniqueName: \"kubernetes.io/projected/1b45e505-3940-45d0-868e-604f8490f8d2-kube-api-access-lc7fr\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.062361 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.062431 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a720d879-8b09-452d-8371-9bf75f28cbde-lock\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.062475 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.062501 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntrtf\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-kube-api-access-ntrtf\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.062527 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a720d879-8b09-452d-8371-9bf75f28cbde-cache\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:44 crc kubenswrapper[4931]: E1129 00:17:44.062696 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:17:44 crc kubenswrapper[4931]: E1129 00:17:44.062733 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 29 00:17:44 crc kubenswrapper[4931]: E1129 00:17:44.062793 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift podName:a720d879-8b09-452d-8371-9bf75f28cbde nodeName:}" failed. No retries permitted until 2025-11-29 00:17:44.562771454 +0000 UTC m=+1077.724664686 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift") pod "swift-storage-0" (UID: "a720d879-8b09-452d-8371-9bf75f28cbde") : configmap "swift-ring-files" not found Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.062709 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/swift-storage-0" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.062972 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a720d879-8b09-452d-8371-9bf75f28cbde-lock\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.063032 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a720d879-8b09-452d-8371-9bf75f28cbde-cache\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.081794 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntrtf\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-kube-api-access-ntrtf\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.099926 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.347748 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-jnbw9"] Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.349179 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.351505 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.352583 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.356315 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.392221 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jnbw9"] Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.398104 4931 generic.go:334] "Generic (PLEG): container finished" podID="1f406c3b-dbb1-421e-9cee-334414d6089f" containerID="6013780e650060493eb5a0ffaf010491e43abc96c9a37d3a49a92e3a0a53fdf6" exitCode=0 Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.398194 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" event={"ID":"1f406c3b-dbb1-421e-9cee-334414d6089f","Type":"ContainerDied","Data":"6013780e650060493eb5a0ffaf010491e43abc96c9a37d3a49a92e3a0a53fdf6"} Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.398217 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" event={"ID":"1f406c3b-dbb1-421e-9cee-334414d6089f","Type":"ContainerStarted","Data":"4a2764f5bc2907df2f53bd86608548af2afe57f2ab04b6dd9999403af860c296"} Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.410785 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-jnbw9"] Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.411819 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-cq95h" event={"ID":"1b45e505-3940-45d0-868e-604f8490f8d2","Type":"ContainerDied","Data":"3a2167d0e06b36233a70da9bc3aa6f18824e70ab174ff5c4af99257213480794"} Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.411870 4931 scope.go:117] "RemoveContainer" containerID="61e39270fc93f60ec85d193da8dd4a5da50fa24061071ba03da712dcd31bd678" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.411891 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-cq95h" Nov 29 00:17:44 crc kubenswrapper[4931]: E1129 00:17:44.418062 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-bphvs ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-jnbw9" podUID="4047ae3f-c8fa-409f-97c8-a65e9d32f746" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.433340 4931 scope.go:117] "RemoveContainer" containerID="83f041170c695791d74550dbca6f2c17a1328731a80329064eef3276919fd45e" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.460555 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-cq95h"] Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.466661 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-cq95h"] Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.469570 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4047ae3f-c8fa-409f-97c8-a65e9d32f746-ring-data-devices\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.469608 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4047ae3f-c8fa-409f-97c8-a65e9d32f746-scripts\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.469656 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bphvs\" (UniqueName: \"kubernetes.io/projected/4047ae3f-c8fa-409f-97c8-a65e9d32f746-kube-api-access-bphvs\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.469743 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-combined-ca-bundle\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.470052 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4047ae3f-c8fa-409f-97c8-a65e9d32f746-etc-swift\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.470091 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-dispersionconf\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.470265 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-swiftconf\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.571300 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4047ae3f-c8fa-409f-97c8-a65e9d32f746-etc-swift\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.571528 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-dispersionconf\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.571625 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4047ae3f-c8fa-409f-97c8-a65e9d32f746-etc-swift\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.571741 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-swiftconf\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.571890 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.571986 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4047ae3f-c8fa-409f-97c8-a65e9d32f746-ring-data-devices\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.572079 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4047ae3f-c8fa-409f-97c8-a65e9d32f746-scripts\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.572184 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bphvs\" (UniqueName: \"kubernetes.io/projected/4047ae3f-c8fa-409f-97c8-a65e9d32f746-kube-api-access-bphvs\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.572284 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-combined-ca-bundle\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: E1129 00:17:44.572125 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:17:44 crc kubenswrapper[4931]: E1129 00:17:44.572505 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 29 00:17:44 crc kubenswrapper[4931]: E1129 00:17:44.572550 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift podName:a720d879-8b09-452d-8371-9bf75f28cbde nodeName:}" failed. No retries permitted until 2025-11-29 00:17:45.572533178 +0000 UTC m=+1078.734426410 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift") pod "swift-storage-0" (UID: "a720d879-8b09-452d-8371-9bf75f28cbde") : configmap "swift-ring-files" not found Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.573245 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4047ae3f-c8fa-409f-97c8-a65e9d32f746-ring-data-devices\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.575264 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-swiftconf\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.575285 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4047ae3f-c8fa-409f-97c8-a65e9d32f746-scripts\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.577108 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-dispersionconf\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.577955 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-combined-ca-bundle\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:44 crc kubenswrapper[4931]: I1129 00:17:44.593173 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bphvs\" (UniqueName: \"kubernetes.io/projected/4047ae3f-c8fa-409f-97c8-a65e9d32f746-kube-api-access-bphvs\") pod \"swift-ring-rebalance-jnbw9\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.228861 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b45e505-3940-45d0-868e-604f8490f8d2" path="/var/lib/kubelet/pods/1b45e505-3940-45d0-868e-604f8490f8d2/volumes" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.426510 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.426606 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" event={"ID":"1f406c3b-dbb1-421e-9cee-334414d6089f","Type":"ContainerStarted","Data":"35db0a6c7af187c4ddc94c6834c37556eaf379e1acccb939989c780e040403c1"} Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.427080 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.448000 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" podStartSLOduration=3.447983904 podStartE2EDuration="3.447983904s" podCreationTimestamp="2025-11-29 00:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:17:45.447884121 +0000 UTC m=+1078.609777383" watchObservedRunningTime="2025-11-29 00:17:45.447983904 +0000 UTC m=+1078.609877136" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.449488 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.587525 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-dispersionconf\") pod \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.587671 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-combined-ca-bundle\") pod \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.587776 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4047ae3f-c8fa-409f-97c8-a65e9d32f746-scripts\") pod \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.587800 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4047ae3f-c8fa-409f-97c8-a65e9d32f746-ring-data-devices\") pod \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.587902 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bphvs\" (UniqueName: \"kubernetes.io/projected/4047ae3f-c8fa-409f-97c8-a65e9d32f746-kube-api-access-bphvs\") pod \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.587941 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4047ae3f-c8fa-409f-97c8-a65e9d32f746-etc-swift\") pod \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.587963 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-swiftconf\") pod \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\" (UID: \"4047ae3f-c8fa-409f-97c8-a65e9d32f746\") " Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.588304 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.588309 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4047ae3f-c8fa-409f-97c8-a65e9d32f746-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "4047ae3f-c8fa-409f-97c8-a65e9d32f746" (UID: "4047ae3f-c8fa-409f-97c8-a65e9d32f746"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.588443 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4047ae3f-c8fa-409f-97c8-a65e9d32f746-scripts" (OuterVolumeSpecName: "scripts") pod "4047ae3f-c8fa-409f-97c8-a65e9d32f746" (UID: "4047ae3f-c8fa-409f-97c8-a65e9d32f746"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:45 crc kubenswrapper[4931]: E1129 00:17:45.588536 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:17:45 crc kubenswrapper[4931]: E1129 00:17:45.588638 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 29 00:17:45 crc kubenswrapper[4931]: E1129 00:17:45.588673 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift podName:a720d879-8b09-452d-8371-9bf75f28cbde nodeName:}" failed. No retries permitted until 2025-11-29 00:17:47.588658971 +0000 UTC m=+1080.750552203 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift") pod "swift-storage-0" (UID: "a720d879-8b09-452d-8371-9bf75f28cbde") : configmap "swift-ring-files" not found Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.589161 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4047ae3f-c8fa-409f-97c8-a65e9d32f746-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "4047ae3f-c8fa-409f-97c8-a65e9d32f746" (UID: "4047ae3f-c8fa-409f-97c8-a65e9d32f746"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.593004 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "4047ae3f-c8fa-409f-97c8-a65e9d32f746" (UID: "4047ae3f-c8fa-409f-97c8-a65e9d32f746"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.593457 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "4047ae3f-c8fa-409f-97c8-a65e9d32f746" (UID: "4047ae3f-c8fa-409f-97c8-a65e9d32f746"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.593456 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4047ae3f-c8fa-409f-97c8-a65e9d32f746-kube-api-access-bphvs" (OuterVolumeSpecName: "kube-api-access-bphvs") pod "4047ae3f-c8fa-409f-97c8-a65e9d32f746" (UID: "4047ae3f-c8fa-409f-97c8-a65e9d32f746"). InnerVolumeSpecName "kube-api-access-bphvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.598044 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4047ae3f-c8fa-409f-97c8-a65e9d32f746" (UID: "4047ae3f-c8fa-409f-97c8-a65e9d32f746"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.690138 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bphvs\" (UniqueName: \"kubernetes.io/projected/4047ae3f-c8fa-409f-97c8-a65e9d32f746-kube-api-access-bphvs\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.690175 4931 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4047ae3f-c8fa-409f-97c8-a65e9d32f746-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.690186 4931 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.690194 4931 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.690203 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4047ae3f-c8fa-409f-97c8-a65e9d32f746-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.690211 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4047ae3f-c8fa-409f-97c8-a65e9d32f746-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:45 crc kubenswrapper[4931]: I1129 00:17:45.690218 4931 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4047ae3f-c8fa-409f-97c8-a65e9d32f746-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.212443 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-9f3d-account-create-update-rfwpv"] Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.213828 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9f3d-account-create-update-rfwpv" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.220332 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-6dghl"] Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.221616 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6dghl" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.228784 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.235098 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9f3d-account-create-update-rfwpv"] Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.254544 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6dghl"] Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.304902 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qldsm\" (UniqueName: \"kubernetes.io/projected/3eb2f135-dcd7-4280-afa3-8f3fd1201b05-kube-api-access-qldsm\") pod \"glance-db-create-6dghl\" (UID: \"3eb2f135-dcd7-4280-afa3-8f3fd1201b05\") " pod="openstack/glance-db-create-6dghl" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.305183 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3eb2f135-dcd7-4280-afa3-8f3fd1201b05-operator-scripts\") pod \"glance-db-create-6dghl\" (UID: \"3eb2f135-dcd7-4280-afa3-8f3fd1201b05\") " pod="openstack/glance-db-create-6dghl" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.305353 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/269a3ba4-671b-47fd-8d81-960cc79810cb-operator-scripts\") pod \"glance-9f3d-account-create-update-rfwpv\" (UID: \"269a3ba4-671b-47fd-8d81-960cc79810cb\") " pod="openstack/glance-9f3d-account-create-update-rfwpv" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.305579 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9n8p\" (UniqueName: \"kubernetes.io/projected/269a3ba4-671b-47fd-8d81-960cc79810cb-kube-api-access-t9n8p\") pod \"glance-9f3d-account-create-update-rfwpv\" (UID: \"269a3ba4-671b-47fd-8d81-960cc79810cb\") " pod="openstack/glance-9f3d-account-create-update-rfwpv" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.406565 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9n8p\" (UniqueName: \"kubernetes.io/projected/269a3ba4-671b-47fd-8d81-960cc79810cb-kube-api-access-t9n8p\") pod \"glance-9f3d-account-create-update-rfwpv\" (UID: \"269a3ba4-671b-47fd-8d81-960cc79810cb\") " pod="openstack/glance-9f3d-account-create-update-rfwpv" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.406693 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qldsm\" (UniqueName: \"kubernetes.io/projected/3eb2f135-dcd7-4280-afa3-8f3fd1201b05-kube-api-access-qldsm\") pod \"glance-db-create-6dghl\" (UID: \"3eb2f135-dcd7-4280-afa3-8f3fd1201b05\") " pod="openstack/glance-db-create-6dghl" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.406723 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3eb2f135-dcd7-4280-afa3-8f3fd1201b05-operator-scripts\") pod \"glance-db-create-6dghl\" (UID: \"3eb2f135-dcd7-4280-afa3-8f3fd1201b05\") " pod="openstack/glance-db-create-6dghl" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.406790 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/269a3ba4-671b-47fd-8d81-960cc79810cb-operator-scripts\") pod \"glance-9f3d-account-create-update-rfwpv\" (UID: \"269a3ba4-671b-47fd-8d81-960cc79810cb\") " pod="openstack/glance-9f3d-account-create-update-rfwpv" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.407662 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/269a3ba4-671b-47fd-8d81-960cc79810cb-operator-scripts\") pod \"glance-9f3d-account-create-update-rfwpv\" (UID: \"269a3ba4-671b-47fd-8d81-960cc79810cb\") " pod="openstack/glance-9f3d-account-create-update-rfwpv" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.407780 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3eb2f135-dcd7-4280-afa3-8f3fd1201b05-operator-scripts\") pod \"glance-db-create-6dghl\" (UID: \"3eb2f135-dcd7-4280-afa3-8f3fd1201b05\") " pod="openstack/glance-db-create-6dghl" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.429764 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qldsm\" (UniqueName: \"kubernetes.io/projected/3eb2f135-dcd7-4280-afa3-8f3fd1201b05-kube-api-access-qldsm\") pod \"glance-db-create-6dghl\" (UID: \"3eb2f135-dcd7-4280-afa3-8f3fd1201b05\") " pod="openstack/glance-db-create-6dghl" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.430063 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9n8p\" (UniqueName: \"kubernetes.io/projected/269a3ba4-671b-47fd-8d81-960cc79810cb-kube-api-access-t9n8p\") pod \"glance-9f3d-account-create-update-rfwpv\" (UID: \"269a3ba4-671b-47fd-8d81-960cc79810cb\") " pod="openstack/glance-9f3d-account-create-update-rfwpv" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.434636 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jnbw9" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.508299 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-jnbw9"] Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.526126 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-jnbw9"] Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.533706 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9f3d-account-create-update-rfwpv" Nov 29 00:17:46 crc kubenswrapper[4931]: I1129 00:17:46.542218 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6dghl" Nov 29 00:17:47 crc kubenswrapper[4931]: I1129 00:17:47.023115 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6dghl"] Nov 29 00:17:47 crc kubenswrapper[4931]: I1129 00:17:47.163851 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9f3d-account-create-update-rfwpv"] Nov 29 00:17:47 crc kubenswrapper[4931]: W1129 00:17:47.169153 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod269a3ba4_671b_47fd_8d81_960cc79810cb.slice/crio-5b4f6bb5331eab99ac5b4907e56f0da1ab881088708f6f5922a21d3a1162e84e WatchSource:0}: Error finding container 5b4f6bb5331eab99ac5b4907e56f0da1ab881088708f6f5922a21d3a1162e84e: Status 404 returned error can't find the container with id 5b4f6bb5331eab99ac5b4907e56f0da1ab881088708f6f5922a21d3a1162e84e Nov 29 00:17:47 crc kubenswrapper[4931]: I1129 00:17:47.175663 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 29 00:17:47 crc kubenswrapper[4931]: I1129 00:17:47.231857 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4047ae3f-c8fa-409f-97c8-a65e9d32f746" path="/var/lib/kubelet/pods/4047ae3f-c8fa-409f-97c8-a65e9d32f746/volumes" Nov 29 00:17:47 crc kubenswrapper[4931]: I1129 00:17:47.441723 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9f3d-account-create-update-rfwpv" event={"ID":"269a3ba4-671b-47fd-8d81-960cc79810cb","Type":"ContainerStarted","Data":"03ae9252b88462a5986482396100bb757d37d0fc08ab3d06303c511433c81ce0"} Nov 29 00:17:47 crc kubenswrapper[4931]: I1129 00:17:47.441773 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9f3d-account-create-update-rfwpv" event={"ID":"269a3ba4-671b-47fd-8d81-960cc79810cb","Type":"ContainerStarted","Data":"5b4f6bb5331eab99ac5b4907e56f0da1ab881088708f6f5922a21d3a1162e84e"} Nov 29 00:17:47 crc kubenswrapper[4931]: I1129 00:17:47.444555 4931 generic.go:334] "Generic (PLEG): container finished" podID="3eb2f135-dcd7-4280-afa3-8f3fd1201b05" containerID="a8945aa397a1bede424f2dde6d339fb659a0295784a089857ee21da174e9ce76" exitCode=0 Nov 29 00:17:47 crc kubenswrapper[4931]: I1129 00:17:47.444648 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6dghl" event={"ID":"3eb2f135-dcd7-4280-afa3-8f3fd1201b05","Type":"ContainerDied","Data":"a8945aa397a1bede424f2dde6d339fb659a0295784a089857ee21da174e9ce76"} Nov 29 00:17:47 crc kubenswrapper[4931]: I1129 00:17:47.444681 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6dghl" event={"ID":"3eb2f135-dcd7-4280-afa3-8f3fd1201b05","Type":"ContainerStarted","Data":"58b5346fd0dd07e7aeb77b53f1c5f0ea36f8671810c38858b6bfa8d97105791e"} Nov 29 00:17:47 crc kubenswrapper[4931]: I1129 00:17:47.461256 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-9f3d-account-create-update-rfwpv" podStartSLOduration=1.461238773 podStartE2EDuration="1.461238773s" podCreationTimestamp="2025-11-29 00:17:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:17:47.459266958 +0000 UTC m=+1080.621160190" watchObservedRunningTime="2025-11-29 00:17:47.461238773 +0000 UTC m=+1080.623132025" Nov 29 00:17:47 crc kubenswrapper[4931]: I1129 00:17:47.628370 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:47 crc kubenswrapper[4931]: E1129 00:17:47.628590 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:17:47 crc kubenswrapper[4931]: E1129 00:17:47.628867 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 29 00:17:47 crc kubenswrapper[4931]: E1129 00:17:47.628933 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift podName:a720d879-8b09-452d-8371-9bf75f28cbde nodeName:}" failed. No retries permitted until 2025-11-29 00:17:51.628912759 +0000 UTC m=+1084.790805991 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift") pod "swift-storage-0" (UID: "a720d879-8b09-452d-8371-9bf75f28cbde") : configmap "swift-ring-files" not found Nov 29 00:17:48 crc kubenswrapper[4931]: I1129 00:17:48.462962 4931 generic.go:334] "Generic (PLEG): container finished" podID="269a3ba4-671b-47fd-8d81-960cc79810cb" containerID="03ae9252b88462a5986482396100bb757d37d0fc08ab3d06303c511433c81ce0" exitCode=0 Nov 29 00:17:48 crc kubenswrapper[4931]: I1129 00:17:48.463392 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9f3d-account-create-update-rfwpv" event={"ID":"269a3ba4-671b-47fd-8d81-960cc79810cb","Type":"ContainerDied","Data":"03ae9252b88462a5986482396100bb757d37d0fc08ab3d06303c511433c81ce0"} Nov 29 00:17:48 crc kubenswrapper[4931]: I1129 00:17:48.863833 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6dghl" Nov 29 00:17:48 crc kubenswrapper[4931]: I1129 00:17:48.949644 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qldsm\" (UniqueName: \"kubernetes.io/projected/3eb2f135-dcd7-4280-afa3-8f3fd1201b05-kube-api-access-qldsm\") pod \"3eb2f135-dcd7-4280-afa3-8f3fd1201b05\" (UID: \"3eb2f135-dcd7-4280-afa3-8f3fd1201b05\") " Nov 29 00:17:48 crc kubenswrapper[4931]: I1129 00:17:48.949754 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3eb2f135-dcd7-4280-afa3-8f3fd1201b05-operator-scripts\") pod \"3eb2f135-dcd7-4280-afa3-8f3fd1201b05\" (UID: \"3eb2f135-dcd7-4280-afa3-8f3fd1201b05\") " Nov 29 00:17:48 crc kubenswrapper[4931]: I1129 00:17:48.951133 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3eb2f135-dcd7-4280-afa3-8f3fd1201b05-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3eb2f135-dcd7-4280-afa3-8f3fd1201b05" (UID: "3eb2f135-dcd7-4280-afa3-8f3fd1201b05"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:48 crc kubenswrapper[4931]: I1129 00:17:48.955989 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3eb2f135-dcd7-4280-afa3-8f3fd1201b05-kube-api-access-qldsm" (OuterVolumeSpecName: "kube-api-access-qldsm") pod "3eb2f135-dcd7-4280-afa3-8f3fd1201b05" (UID: "3eb2f135-dcd7-4280-afa3-8f3fd1201b05"). InnerVolumeSpecName "kube-api-access-qldsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:49 crc kubenswrapper[4931]: I1129 00:17:49.052430 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qldsm\" (UniqueName: \"kubernetes.io/projected/3eb2f135-dcd7-4280-afa3-8f3fd1201b05-kube-api-access-qldsm\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:49 crc kubenswrapper[4931]: I1129 00:17:49.052473 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3eb2f135-dcd7-4280-afa3-8f3fd1201b05-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:49 crc kubenswrapper[4931]: I1129 00:17:49.472322 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6dghl" Nov 29 00:17:49 crc kubenswrapper[4931]: I1129 00:17:49.472322 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6dghl" event={"ID":"3eb2f135-dcd7-4280-afa3-8f3fd1201b05","Type":"ContainerDied","Data":"58b5346fd0dd07e7aeb77b53f1c5f0ea36f8671810c38858b6bfa8d97105791e"} Nov 29 00:17:49 crc kubenswrapper[4931]: I1129 00:17:49.472407 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58b5346fd0dd07e7aeb77b53f1c5f0ea36f8671810c38858b6bfa8d97105791e" Nov 29 00:17:49 crc kubenswrapper[4931]: I1129 00:17:49.779676 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9f3d-account-create-update-rfwpv" Nov 29 00:17:49 crc kubenswrapper[4931]: I1129 00:17:49.864055 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/269a3ba4-671b-47fd-8d81-960cc79810cb-operator-scripts\") pod \"269a3ba4-671b-47fd-8d81-960cc79810cb\" (UID: \"269a3ba4-671b-47fd-8d81-960cc79810cb\") " Nov 29 00:17:49 crc kubenswrapper[4931]: I1129 00:17:49.864295 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9n8p\" (UniqueName: \"kubernetes.io/projected/269a3ba4-671b-47fd-8d81-960cc79810cb-kube-api-access-t9n8p\") pod \"269a3ba4-671b-47fd-8d81-960cc79810cb\" (UID: \"269a3ba4-671b-47fd-8d81-960cc79810cb\") " Nov 29 00:17:49 crc kubenswrapper[4931]: I1129 00:17:49.864827 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/269a3ba4-671b-47fd-8d81-960cc79810cb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "269a3ba4-671b-47fd-8d81-960cc79810cb" (UID: "269a3ba4-671b-47fd-8d81-960cc79810cb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:49 crc kubenswrapper[4931]: I1129 00:17:49.870507 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/269a3ba4-671b-47fd-8d81-960cc79810cb-kube-api-access-t9n8p" (OuterVolumeSpecName: "kube-api-access-t9n8p") pod "269a3ba4-671b-47fd-8d81-960cc79810cb" (UID: "269a3ba4-671b-47fd-8d81-960cc79810cb"). InnerVolumeSpecName "kube-api-access-t9n8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:49 crc kubenswrapper[4931]: I1129 00:17:49.966414 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/269a3ba4-671b-47fd-8d81-960cc79810cb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:49 crc kubenswrapper[4931]: I1129 00:17:49.966446 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9n8p\" (UniqueName: \"kubernetes.io/projected/269a3ba4-671b-47fd-8d81-960cc79810cb-kube-api-access-t9n8p\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.474790 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-f7mvr"] Nov 29 00:17:50 crc kubenswrapper[4931]: E1129 00:17:50.475571 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269a3ba4-671b-47fd-8d81-960cc79810cb" containerName="mariadb-account-create-update" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.475596 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="269a3ba4-671b-47fd-8d81-960cc79810cb" containerName="mariadb-account-create-update" Nov 29 00:17:50 crc kubenswrapper[4931]: E1129 00:17:50.475634 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eb2f135-dcd7-4280-afa3-8f3fd1201b05" containerName="mariadb-database-create" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.475645 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eb2f135-dcd7-4280-afa3-8f3fd1201b05" containerName="mariadb-database-create" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.475924 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="3eb2f135-dcd7-4280-afa3-8f3fd1201b05" containerName="mariadb-database-create" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.475970 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="269a3ba4-671b-47fd-8d81-960cc79810cb" containerName="mariadb-account-create-update" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.476689 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-f7mvr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.485401 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-f7mvr"] Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.485572 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9f3d-account-create-update-rfwpv" event={"ID":"269a3ba4-671b-47fd-8d81-960cc79810cb","Type":"ContainerDied","Data":"5b4f6bb5331eab99ac5b4907e56f0da1ab881088708f6f5922a21d3a1162e84e"} Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.485612 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b4f6bb5331eab99ac5b4907e56f0da1ab881088708f6f5922a21d3a1162e84e" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.485685 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9f3d-account-create-update-rfwpv" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.575609 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8l47\" (UniqueName: \"kubernetes.io/projected/0032d91c-6fc0-4f0a-8410-ccf56967c6f8-kube-api-access-h8l47\") pod \"keystone-db-create-f7mvr\" (UID: \"0032d91c-6fc0-4f0a-8410-ccf56967c6f8\") " pod="openstack/keystone-db-create-f7mvr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.575691 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0032d91c-6fc0-4f0a-8410-ccf56967c6f8-operator-scripts\") pod \"keystone-db-create-f7mvr\" (UID: \"0032d91c-6fc0-4f0a-8410-ccf56967c6f8\") " pod="openstack/keystone-db-create-f7mvr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.593600 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-ab0a-account-create-update-cjkxg"] Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.594970 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ab0a-account-create-update-cjkxg" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.597465 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.624982 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ab0a-account-create-update-cjkxg"] Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.678657 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0032d91c-6fc0-4f0a-8410-ccf56967c6f8-operator-scripts\") pod \"keystone-db-create-f7mvr\" (UID: \"0032d91c-6fc0-4f0a-8410-ccf56967c6f8\") " pod="openstack/keystone-db-create-f7mvr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.678734 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf-operator-scripts\") pod \"keystone-ab0a-account-create-update-cjkxg\" (UID: \"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf\") " pod="openstack/keystone-ab0a-account-create-update-cjkxg" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.678882 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4686p\" (UniqueName: \"kubernetes.io/projected/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf-kube-api-access-4686p\") pod \"keystone-ab0a-account-create-update-cjkxg\" (UID: \"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf\") " pod="openstack/keystone-ab0a-account-create-update-cjkxg" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.679035 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8l47\" (UniqueName: \"kubernetes.io/projected/0032d91c-6fc0-4f0a-8410-ccf56967c6f8-kube-api-access-h8l47\") pod \"keystone-db-create-f7mvr\" (UID: \"0032d91c-6fc0-4f0a-8410-ccf56967c6f8\") " pod="openstack/keystone-db-create-f7mvr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.679386 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0032d91c-6fc0-4f0a-8410-ccf56967c6f8-operator-scripts\") pod \"keystone-db-create-f7mvr\" (UID: \"0032d91c-6fc0-4f0a-8410-ccf56967c6f8\") " pod="openstack/keystone-db-create-f7mvr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.698937 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8l47\" (UniqueName: \"kubernetes.io/projected/0032d91c-6fc0-4f0a-8410-ccf56967c6f8-kube-api-access-h8l47\") pod \"keystone-db-create-f7mvr\" (UID: \"0032d91c-6fc0-4f0a-8410-ccf56967c6f8\") " pod="openstack/keystone-db-create-f7mvr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.780995 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf-operator-scripts\") pod \"keystone-ab0a-account-create-update-cjkxg\" (UID: \"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf\") " pod="openstack/keystone-ab0a-account-create-update-cjkxg" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.781081 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4686p\" (UniqueName: \"kubernetes.io/projected/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf-kube-api-access-4686p\") pod \"keystone-ab0a-account-create-update-cjkxg\" (UID: \"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf\") " pod="openstack/keystone-ab0a-account-create-update-cjkxg" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.781728 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf-operator-scripts\") pod \"keystone-ab0a-account-create-update-cjkxg\" (UID: \"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf\") " pod="openstack/keystone-ab0a-account-create-update-cjkxg" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.799388 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-f7mvr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.799467 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4686p\" (UniqueName: \"kubernetes.io/projected/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf-kube-api-access-4686p\") pod \"keystone-ab0a-account-create-update-cjkxg\" (UID: \"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf\") " pod="openstack/keystone-ab0a-account-create-update-cjkxg" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.817921 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-n4xfr"] Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.819156 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-n4xfr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.838737 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-n4xfr"] Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.882662 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e-operator-scripts\") pod \"placement-db-create-n4xfr\" (UID: \"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e\") " pod="openstack/placement-db-create-n4xfr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.882735 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7mmr\" (UniqueName: \"kubernetes.io/projected/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e-kube-api-access-k7mmr\") pod \"placement-db-create-n4xfr\" (UID: \"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e\") " pod="openstack/placement-db-create-n4xfr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.926360 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ab0a-account-create-update-cjkxg" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.929852 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-ed60-account-create-update-9thjn"] Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.932615 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ed60-account-create-update-9thjn" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.937916 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.955176 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.967843 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ed60-account-create-update-9thjn"] Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.987519 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e-operator-scripts\") pod \"placement-db-create-n4xfr\" (UID: \"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e\") " pod="openstack/placement-db-create-n4xfr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.987569 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7mmr\" (UniqueName: \"kubernetes.io/projected/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e-kube-api-access-k7mmr\") pod \"placement-db-create-n4xfr\" (UID: \"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e\") " pod="openstack/placement-db-create-n4xfr" Nov 29 00:17:50 crc kubenswrapper[4931]: I1129 00:17:50.988428 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e-operator-scripts\") pod \"placement-db-create-n4xfr\" (UID: \"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e\") " pod="openstack/placement-db-create-n4xfr" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.017420 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7mmr\" (UniqueName: \"kubernetes.io/projected/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e-kube-api-access-k7mmr\") pod \"placement-db-create-n4xfr\" (UID: \"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e\") " pod="openstack/placement-db-create-n4xfr" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.089073 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f55707a4-27e4-496f-ae85-1e07ddf91a5b-operator-scripts\") pod \"placement-ed60-account-create-update-9thjn\" (UID: \"f55707a4-27e4-496f-ae85-1e07ddf91a5b\") " pod="openstack/placement-ed60-account-create-update-9thjn" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.089860 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwmr7\" (UniqueName: \"kubernetes.io/projected/f55707a4-27e4-496f-ae85-1e07ddf91a5b-kube-api-access-nwmr7\") pod \"placement-ed60-account-create-update-9thjn\" (UID: \"f55707a4-27e4-496f-ae85-1e07ddf91a5b\") " pod="openstack/placement-ed60-account-create-update-9thjn" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.191523 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-n4xfr" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.191922 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f55707a4-27e4-496f-ae85-1e07ddf91a5b-operator-scripts\") pod \"placement-ed60-account-create-update-9thjn\" (UID: \"f55707a4-27e4-496f-ae85-1e07ddf91a5b\") " pod="openstack/placement-ed60-account-create-update-9thjn" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.192058 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwmr7\" (UniqueName: \"kubernetes.io/projected/f55707a4-27e4-496f-ae85-1e07ddf91a5b-kube-api-access-nwmr7\") pod \"placement-ed60-account-create-update-9thjn\" (UID: \"f55707a4-27e4-496f-ae85-1e07ddf91a5b\") " pod="openstack/placement-ed60-account-create-update-9thjn" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.192649 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f55707a4-27e4-496f-ae85-1e07ddf91a5b-operator-scripts\") pod \"placement-ed60-account-create-update-9thjn\" (UID: \"f55707a4-27e4-496f-ae85-1e07ddf91a5b\") " pod="openstack/placement-ed60-account-create-update-9thjn" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.217415 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwmr7\" (UniqueName: \"kubernetes.io/projected/f55707a4-27e4-496f-ae85-1e07ddf91a5b-kube-api-access-nwmr7\") pod \"placement-ed60-account-create-update-9thjn\" (UID: \"f55707a4-27e4-496f-ae85-1e07ddf91a5b\") " pod="openstack/placement-ed60-account-create-update-9thjn" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.292310 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-f7mvr"] Nov 29 00:17:51 crc kubenswrapper[4931]: W1129 00:17:51.303134 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0032d91c_6fc0_4f0a_8410_ccf56967c6f8.slice/crio-b08679ff58c9cb126f21f92b995a7361c79a84db883c292d306afee99010d3cd WatchSource:0}: Error finding container b08679ff58c9cb126f21f92b995a7361c79a84db883c292d306afee99010d3cd: Status 404 returned error can't find the container with id b08679ff58c9cb126f21f92b995a7361c79a84db883c292d306afee99010d3cd Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.371605 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ed60-account-create-update-9thjn" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.433097 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ab0a-account-create-update-cjkxg"] Nov 29 00:17:51 crc kubenswrapper[4931]: W1129 00:17:51.454327 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7a6941c_cc1c_4b2b_89f4_b6b866ef2abf.slice/crio-d840e32bd6e93cce7fc0a9c188bbdfc3030c367662e2301571f8e31e6ac6f4e0 WatchSource:0}: Error finding container d840e32bd6e93cce7fc0a9c188bbdfc3030c367662e2301571f8e31e6ac6f4e0: Status 404 returned error can't find the container with id d840e32bd6e93cce7fc0a9c188bbdfc3030c367662e2301571f8e31e6ac6f4e0 Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.505015 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-f7mvr" event={"ID":"0032d91c-6fc0-4f0a-8410-ccf56967c6f8","Type":"ContainerStarted","Data":"b08679ff58c9cb126f21f92b995a7361c79a84db883c292d306afee99010d3cd"} Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.507522 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ab0a-account-create-update-cjkxg" event={"ID":"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf","Type":"ContainerStarted","Data":"d840e32bd6e93cce7fc0a9c188bbdfc3030c367662e2301571f8e31e6ac6f4e0"} Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.533516 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-dj2pb"] Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.534796 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.537077 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-m5lx4" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.540391 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.557944 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-dj2pb"] Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.602941 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-config-data\") pod \"glance-db-sync-dj2pb\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.602993 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-825k8\" (UniqueName: \"kubernetes.io/projected/2c957f2c-45cc-4a11-8019-e20de202266b-kube-api-access-825k8\") pod \"glance-db-sync-dj2pb\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.603024 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-db-sync-config-data\") pod \"glance-db-sync-dj2pb\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.603068 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-combined-ca-bundle\") pod \"glance-db-sync-dj2pb\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: W1129 00:17:51.663302 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7c9f10f_3823_4af5_9f94_ee8c5ef79d2e.slice/crio-307d463188e8ad2f9422c0265a5ee6d25b9556b2b3086ce0f460df6fa4434ec1 WatchSource:0}: Error finding container 307d463188e8ad2f9422c0265a5ee6d25b9556b2b3086ce0f460df6fa4434ec1: Status 404 returned error can't find the container with id 307d463188e8ad2f9422c0265a5ee6d25b9556b2b3086ce0f460df6fa4434ec1 Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.668594 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-n4xfr"] Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.704384 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.704454 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-config-data\") pod \"glance-db-sync-dj2pb\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.704486 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-825k8\" (UniqueName: \"kubernetes.io/projected/2c957f2c-45cc-4a11-8019-e20de202266b-kube-api-access-825k8\") pod \"glance-db-sync-dj2pb\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.704520 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-db-sync-config-data\") pod \"glance-db-sync-dj2pb\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.704584 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-combined-ca-bundle\") pod \"glance-db-sync-dj2pb\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: E1129 00:17:51.704578 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:17:51 crc kubenswrapper[4931]: E1129 00:17:51.704663 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 29 00:17:51 crc kubenswrapper[4931]: E1129 00:17:51.704719 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift podName:a720d879-8b09-452d-8371-9bf75f28cbde nodeName:}" failed. No retries permitted until 2025-11-29 00:17:59.704702004 +0000 UTC m=+1092.866595246 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift") pod "swift-storage-0" (UID: "a720d879-8b09-452d-8371-9bf75f28cbde") : configmap "swift-ring-files" not found Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.806597 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-db-sync-config-data\") pod \"glance-db-sync-dj2pb\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.807929 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-config-data\") pod \"glance-db-sync-dj2pb\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.808106 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-combined-ca-bundle\") pod \"glance-db-sync-dj2pb\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.809262 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-825k8\" (UniqueName: \"kubernetes.io/projected/2c957f2c-45cc-4a11-8019-e20de202266b-kube-api-access-825k8\") pod \"glance-db-sync-dj2pb\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:51 crc kubenswrapper[4931]: W1129 00:17:51.859221 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf55707a4_27e4_496f_ae85_1e07ddf91a5b.slice/crio-caad887c59ef3bfe8ca0340132cae7db5b45c3ea4e2816790f83cd7d2ee838f8 WatchSource:0}: Error finding container caad887c59ef3bfe8ca0340132cae7db5b45c3ea4e2816790f83cd7d2ee838f8: Status 404 returned error can't find the container with id caad887c59ef3bfe8ca0340132cae7db5b45c3ea4e2816790f83cd7d2ee838f8 Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.861716 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ed60-account-create-update-9thjn"] Nov 29 00:17:51 crc kubenswrapper[4931]: I1129 00:17:51.895960 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-dj2pb" Nov 29 00:17:52 crc kubenswrapper[4931]: I1129 00:17:52.464020 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-dj2pb"] Nov 29 00:17:52 crc kubenswrapper[4931]: I1129 00:17:52.515704 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-dj2pb" event={"ID":"2c957f2c-45cc-4a11-8019-e20de202266b","Type":"ContainerStarted","Data":"b01a3e25c1ff1510946ca9a777be650d80a7ef87abda1e49055488f4779af556"} Nov 29 00:17:52 crc kubenswrapper[4931]: I1129 00:17:52.517209 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-n4xfr" event={"ID":"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e","Type":"ContainerStarted","Data":"307d463188e8ad2f9422c0265a5ee6d25b9556b2b3086ce0f460df6fa4434ec1"} Nov 29 00:17:52 crc kubenswrapper[4931]: I1129 00:17:52.518437 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ed60-account-create-update-9thjn" event={"ID":"f55707a4-27e4-496f-ae85-1e07ddf91a5b","Type":"ContainerStarted","Data":"caad887c59ef3bfe8ca0340132cae7db5b45c3ea4e2816790f83cd7d2ee838f8"} Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.083935 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.146028 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8jzz9"] Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.146311 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" podUID="a0b54682-21aa-4382-b7c4-5db01da1cab8" containerName="dnsmasq-dns" containerID="cri-o://813fce3f7ef18e8abaeaf642f1b9fbf0709577ebc61d60170d1c8e78c34790d7" gracePeriod=10 Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.527672 4931 generic.go:334] "Generic (PLEG): container finished" podID="0032d91c-6fc0-4f0a-8410-ccf56967c6f8" containerID="226248938d875e035fa0361171b0b911e810a8bfa3d2c7e6365d02e15a8f3c63" exitCode=0 Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.527929 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-f7mvr" event={"ID":"0032d91c-6fc0-4f0a-8410-ccf56967c6f8","Type":"ContainerDied","Data":"226248938d875e035fa0361171b0b911e810a8bfa3d2c7e6365d02e15a8f3c63"} Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.531444 4931 generic.go:334] "Generic (PLEG): container finished" podID="d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf" containerID="0af4b3337efba61d068d58053b627c4a625de291fd684680f1e0896589628714" exitCode=0 Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.531490 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ab0a-account-create-update-cjkxg" event={"ID":"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf","Type":"ContainerDied","Data":"0af4b3337efba61d068d58053b627c4a625de291fd684680f1e0896589628714"} Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.533275 4931 generic.go:334] "Generic (PLEG): container finished" podID="a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e" containerID="5cd01f9aec53fc4d19fcaeff2d34dd831adf986bd86afa31dd5492e48d08ff66" exitCode=0 Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.533326 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-n4xfr" event={"ID":"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e","Type":"ContainerDied","Data":"5cd01f9aec53fc4d19fcaeff2d34dd831adf986bd86afa31dd5492e48d08ff66"} Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.534798 4931 generic.go:334] "Generic (PLEG): container finished" podID="f55707a4-27e4-496f-ae85-1e07ddf91a5b" containerID="ecaac952158742287c253bb03fdfe5e9c8c2b43a4749cd15a510cd6aac3d42ca" exitCode=0 Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.534879 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ed60-account-create-update-9thjn" event={"ID":"f55707a4-27e4-496f-ae85-1e07ddf91a5b","Type":"ContainerDied","Data":"ecaac952158742287c253bb03fdfe5e9c8c2b43a4749cd15a510cd6aac3d42ca"} Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.540339 4931 generic.go:334] "Generic (PLEG): container finished" podID="a0b54682-21aa-4382-b7c4-5db01da1cab8" containerID="813fce3f7ef18e8abaeaf642f1b9fbf0709577ebc61d60170d1c8e78c34790d7" exitCode=0 Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.540366 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" event={"ID":"a0b54682-21aa-4382-b7c4-5db01da1cab8","Type":"ContainerDied","Data":"813fce3f7ef18e8abaeaf642f1b9fbf0709577ebc61d60170d1c8e78c34790d7"} Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.679362 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.740717 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mr2pk\" (UniqueName: \"kubernetes.io/projected/a0b54682-21aa-4382-b7c4-5db01da1cab8-kube-api-access-mr2pk\") pod \"a0b54682-21aa-4382-b7c4-5db01da1cab8\" (UID: \"a0b54682-21aa-4382-b7c4-5db01da1cab8\") " Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.740837 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a0b54682-21aa-4382-b7c4-5db01da1cab8-dns-svc\") pod \"a0b54682-21aa-4382-b7c4-5db01da1cab8\" (UID: \"a0b54682-21aa-4382-b7c4-5db01da1cab8\") " Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.740936 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b54682-21aa-4382-b7c4-5db01da1cab8-config\") pod \"a0b54682-21aa-4382-b7c4-5db01da1cab8\" (UID: \"a0b54682-21aa-4382-b7c4-5db01da1cab8\") " Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.791522 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0b54682-21aa-4382-b7c4-5db01da1cab8-kube-api-access-mr2pk" (OuterVolumeSpecName: "kube-api-access-mr2pk") pod "a0b54682-21aa-4382-b7c4-5db01da1cab8" (UID: "a0b54682-21aa-4382-b7c4-5db01da1cab8"). InnerVolumeSpecName "kube-api-access-mr2pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.837438 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0b54682-21aa-4382-b7c4-5db01da1cab8-config" (OuterVolumeSpecName: "config") pod "a0b54682-21aa-4382-b7c4-5db01da1cab8" (UID: "a0b54682-21aa-4382-b7c4-5db01da1cab8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.839423 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0b54682-21aa-4382-b7c4-5db01da1cab8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a0b54682-21aa-4382-b7c4-5db01da1cab8" (UID: "a0b54682-21aa-4382-b7c4-5db01da1cab8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.844590 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a0b54682-21aa-4382-b7c4-5db01da1cab8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.844626 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b54682-21aa-4382-b7c4-5db01da1cab8-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:53 crc kubenswrapper[4931]: I1129 00:17:53.844640 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mr2pk\" (UniqueName: \"kubernetes.io/projected/a0b54682-21aa-4382-b7c4-5db01da1cab8-kube-api-access-mr2pk\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:54 crc kubenswrapper[4931]: I1129 00:17:54.550209 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" Nov 29 00:17:54 crc kubenswrapper[4931]: I1129 00:17:54.550225 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8jzz9" event={"ID":"a0b54682-21aa-4382-b7c4-5db01da1cab8","Type":"ContainerDied","Data":"c95d509507dde8b4c99413c0a84d8a5f416aec93749f4f74154f8666b413ef17"} Nov 29 00:17:54 crc kubenswrapper[4931]: I1129 00:17:54.550289 4931 scope.go:117] "RemoveContainer" containerID="813fce3f7ef18e8abaeaf642f1b9fbf0709577ebc61d60170d1c8e78c34790d7" Nov 29 00:17:54 crc kubenswrapper[4931]: I1129 00:17:54.595416 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8jzz9"] Nov 29 00:17:54 crc kubenswrapper[4931]: I1129 00:17:54.598559 4931 scope.go:117] "RemoveContainer" containerID="21e5487bc5744a60090acfcddcd06012d736d258de0c9c771c2aac432a6910e8" Nov 29 00:17:54 crc kubenswrapper[4931]: I1129 00:17:54.602892 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8jzz9"] Nov 29 00:17:54 crc kubenswrapper[4931]: I1129 00:17:54.942535 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ed60-account-create-update-9thjn" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.016165 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-n4xfr" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.032276 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-f7mvr" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.067066 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e-operator-scripts\") pod \"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e\" (UID: \"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e\") " Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.067121 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwmr7\" (UniqueName: \"kubernetes.io/projected/f55707a4-27e4-496f-ae85-1e07ddf91a5b-kube-api-access-nwmr7\") pod \"f55707a4-27e4-496f-ae85-1e07ddf91a5b\" (UID: \"f55707a4-27e4-496f-ae85-1e07ddf91a5b\") " Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.067146 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f55707a4-27e4-496f-ae85-1e07ddf91a5b-operator-scripts\") pod \"f55707a4-27e4-496f-ae85-1e07ddf91a5b\" (UID: \"f55707a4-27e4-496f-ae85-1e07ddf91a5b\") " Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.067275 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7mmr\" (UniqueName: \"kubernetes.io/projected/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e-kube-api-access-k7mmr\") pod \"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e\" (UID: \"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e\") " Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.067884 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e" (UID: "a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.067932 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f55707a4-27e4-496f-ae85-1e07ddf91a5b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f55707a4-27e4-496f-ae85-1e07ddf91a5b" (UID: "f55707a4-27e4-496f-ae85-1e07ddf91a5b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.071771 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f55707a4-27e4-496f-ae85-1e07ddf91a5b-kube-api-access-nwmr7" (OuterVolumeSpecName: "kube-api-access-nwmr7") pod "f55707a4-27e4-496f-ae85-1e07ddf91a5b" (UID: "f55707a4-27e4-496f-ae85-1e07ddf91a5b"). InnerVolumeSpecName "kube-api-access-nwmr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.071861 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e-kube-api-access-k7mmr" (OuterVolumeSpecName: "kube-api-access-k7mmr") pod "a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e" (UID: "a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e"). InnerVolumeSpecName "kube-api-access-k7mmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.086571 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ab0a-account-create-update-cjkxg" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.168542 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8l47\" (UniqueName: \"kubernetes.io/projected/0032d91c-6fc0-4f0a-8410-ccf56967c6f8-kube-api-access-h8l47\") pod \"0032d91c-6fc0-4f0a-8410-ccf56967c6f8\" (UID: \"0032d91c-6fc0-4f0a-8410-ccf56967c6f8\") " Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.168621 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf-operator-scripts\") pod \"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf\" (UID: \"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf\") " Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.168763 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0032d91c-6fc0-4f0a-8410-ccf56967c6f8-operator-scripts\") pod \"0032d91c-6fc0-4f0a-8410-ccf56967c6f8\" (UID: \"0032d91c-6fc0-4f0a-8410-ccf56967c6f8\") " Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.168838 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4686p\" (UniqueName: \"kubernetes.io/projected/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf-kube-api-access-4686p\") pod \"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf\" (UID: \"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf\") " Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.169264 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.169291 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwmr7\" (UniqueName: \"kubernetes.io/projected/f55707a4-27e4-496f-ae85-1e07ddf91a5b-kube-api-access-nwmr7\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.169307 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f55707a4-27e4-496f-ae85-1e07ddf91a5b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.169322 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7mmr\" (UniqueName: \"kubernetes.io/projected/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e-kube-api-access-k7mmr\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.169333 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0032d91c-6fc0-4f0a-8410-ccf56967c6f8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0032d91c-6fc0-4f0a-8410-ccf56967c6f8" (UID: "0032d91c-6fc0-4f0a-8410-ccf56967c6f8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.169358 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf" (UID: "d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.171323 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf-kube-api-access-4686p" (OuterVolumeSpecName: "kube-api-access-4686p") pod "d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf" (UID: "d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf"). InnerVolumeSpecName "kube-api-access-4686p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.172307 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0032d91c-6fc0-4f0a-8410-ccf56967c6f8-kube-api-access-h8l47" (OuterVolumeSpecName: "kube-api-access-h8l47") pod "0032d91c-6fc0-4f0a-8410-ccf56967c6f8" (UID: "0032d91c-6fc0-4f0a-8410-ccf56967c6f8"). InnerVolumeSpecName "kube-api-access-h8l47". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.222019 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0b54682-21aa-4382-b7c4-5db01da1cab8" path="/var/lib/kubelet/pods/a0b54682-21aa-4382-b7c4-5db01da1cab8/volumes" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.271182 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8l47\" (UniqueName: \"kubernetes.io/projected/0032d91c-6fc0-4f0a-8410-ccf56967c6f8-kube-api-access-h8l47\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.271606 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.271622 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0032d91c-6fc0-4f0a-8410-ccf56967c6f8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.271635 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4686p\" (UniqueName: \"kubernetes.io/projected/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf-kube-api-access-4686p\") on node \"crc\" DevicePath \"\"" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.561623 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-f7mvr" event={"ID":"0032d91c-6fc0-4f0a-8410-ccf56967c6f8","Type":"ContainerDied","Data":"b08679ff58c9cb126f21f92b995a7361c79a84db883c292d306afee99010d3cd"} Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.562005 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b08679ff58c9cb126f21f92b995a7361c79a84db883c292d306afee99010d3cd" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.561641 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-f7mvr" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.564623 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ab0a-account-create-update-cjkxg" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.564648 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ab0a-account-create-update-cjkxg" event={"ID":"d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf","Type":"ContainerDied","Data":"d840e32bd6e93cce7fc0a9c188bbdfc3030c367662e2301571f8e31e6ac6f4e0"} Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.564675 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d840e32bd6e93cce7fc0a9c188bbdfc3030c367662e2301571f8e31e6ac6f4e0" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.567407 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-n4xfr" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.567431 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-n4xfr" event={"ID":"a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e","Type":"ContainerDied","Data":"307d463188e8ad2f9422c0265a5ee6d25b9556b2b3086ce0f460df6fa4434ec1"} Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.567496 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="307d463188e8ad2f9422c0265a5ee6d25b9556b2b3086ce0f460df6fa4434ec1" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.569497 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ed60-account-create-update-9thjn" Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.569493 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ed60-account-create-update-9thjn" event={"ID":"f55707a4-27e4-496f-ae85-1e07ddf91a5b","Type":"ContainerDied","Data":"caad887c59ef3bfe8ca0340132cae7db5b45c3ea4e2816790f83cd7d2ee838f8"} Nov 29 00:17:55 crc kubenswrapper[4931]: I1129 00:17:55.569651 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caad887c59ef3bfe8ca0340132cae7db5b45c3ea4e2816790f83cd7d2ee838f8" Nov 29 00:17:59 crc kubenswrapper[4931]: I1129 00:17:59.766005 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:17:59 crc kubenswrapper[4931]: E1129 00:17:59.766221 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:17:59 crc kubenswrapper[4931]: E1129 00:17:59.767319 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 29 00:17:59 crc kubenswrapper[4931]: E1129 00:17:59.767386 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift podName:a720d879-8b09-452d-8371-9bf75f28cbde nodeName:}" failed. No retries permitted until 2025-11-29 00:18:15.767368181 +0000 UTC m=+1108.929261433 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift") pod "swift-storage-0" (UID: "a720d879-8b09-452d-8371-9bf75f28cbde") : configmap "swift-ring-files" not found Nov 29 00:18:01 crc kubenswrapper[4931]: I1129 00:18:01.820949 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-trsph" podUID="a64bc6bd-8033-45c9-b2d7-9602aea09225" containerName="ovn-controller" probeResult="failure" output=< Nov 29 00:18:01 crc kubenswrapper[4931]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 29 00:18:01 crc kubenswrapper[4931]: > Nov 29 00:18:01 crc kubenswrapper[4931]: I1129 00:18:01.878732 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:18:01 crc kubenswrapper[4931]: I1129 00:18:01.887226 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7s9fx" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.096805 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-trsph-config-v55dj"] Nov 29 00:18:02 crc kubenswrapper[4931]: E1129 00:18:02.097134 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0032d91c-6fc0-4f0a-8410-ccf56967c6f8" containerName="mariadb-database-create" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.097152 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="0032d91c-6fc0-4f0a-8410-ccf56967c6f8" containerName="mariadb-database-create" Nov 29 00:18:02 crc kubenswrapper[4931]: E1129 00:18:02.097166 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f55707a4-27e4-496f-ae85-1e07ddf91a5b" containerName="mariadb-account-create-update" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.097172 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f55707a4-27e4-496f-ae85-1e07ddf91a5b" containerName="mariadb-account-create-update" Nov 29 00:18:02 crc kubenswrapper[4931]: E1129 00:18:02.097196 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf" containerName="mariadb-account-create-update" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.097204 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf" containerName="mariadb-account-create-update" Nov 29 00:18:02 crc kubenswrapper[4931]: E1129 00:18:02.097218 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b54682-21aa-4382-b7c4-5db01da1cab8" containerName="dnsmasq-dns" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.097224 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b54682-21aa-4382-b7c4-5db01da1cab8" containerName="dnsmasq-dns" Nov 29 00:18:02 crc kubenswrapper[4931]: E1129 00:18:02.097231 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b54682-21aa-4382-b7c4-5db01da1cab8" containerName="init" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.097237 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b54682-21aa-4382-b7c4-5db01da1cab8" containerName="init" Nov 29 00:18:02 crc kubenswrapper[4931]: E1129 00:18:02.097247 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e" containerName="mariadb-database-create" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.097254 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e" containerName="mariadb-database-create" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.097396 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e" containerName="mariadb-database-create" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.097417 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf" containerName="mariadb-account-create-update" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.097427 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b54682-21aa-4382-b7c4-5db01da1cab8" containerName="dnsmasq-dns" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.097437 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f55707a4-27e4-496f-ae85-1e07ddf91a5b" containerName="mariadb-account-create-update" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.097446 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="0032d91c-6fc0-4f0a-8410-ccf56967c6f8" containerName="mariadb-database-create" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.097991 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.100097 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.107975 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-trsph-config-v55dj"] Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.215108 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-run\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.215344 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-log-ovn\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.215475 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbpdf\" (UniqueName: \"kubernetes.io/projected/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-kube-api-access-xbpdf\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.215651 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-scripts\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.215779 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-additional-scripts\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.215894 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-run-ovn\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.317898 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-run-ovn\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.318085 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-run\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.318116 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-log-ovn\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.318150 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbpdf\" (UniqueName: \"kubernetes.io/projected/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-kube-api-access-xbpdf\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.318230 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-run-ovn\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.318265 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-scripts\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.318311 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-additional-scripts\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.318463 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-log-ovn\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.318559 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-run\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.319050 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-additional-scripts\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.320667 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-scripts\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.340851 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbpdf\" (UniqueName: \"kubernetes.io/projected/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-kube-api-access-xbpdf\") pod \"ovn-controller-trsph-config-v55dj\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.418089 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.637936 4931 generic.go:334] "Generic (PLEG): container finished" podID="0d661705-7bb5-482f-b2a8-79b44f6351ac" containerID="c13dfad395ef5b782b6371d50ee00ee1d0b2fdc386adbf7dfeb5c206bbf2019d" exitCode=0 Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.638019 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0d661705-7bb5-482f-b2a8-79b44f6351ac","Type":"ContainerDied","Data":"c13dfad395ef5b782b6371d50ee00ee1d0b2fdc386adbf7dfeb5c206bbf2019d"} Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.642249 4931 generic.go:334] "Generic (PLEG): container finished" podID="fbd78afc-2d6d-4118-92a2-e3499560359f" containerID="7fe4cd6379e10e7b3ca34aac9701c07da875d4602076ce55cb8db09358f686be" exitCode=0 Nov 29 00:18:02 crc kubenswrapper[4931]: I1129 00:18:02.642935 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fbd78afc-2d6d-4118-92a2-e3499560359f","Type":"ContainerDied","Data":"7fe4cd6379e10e7b3ca34aac9701c07da875d4602076ce55cb8db09358f686be"} Nov 29 00:18:06 crc kubenswrapper[4931]: I1129 00:18:06.832868 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-trsph" podUID="a64bc6bd-8033-45c9-b2d7-9602aea09225" containerName="ovn-controller" probeResult="failure" output=< Nov 29 00:18:06 crc kubenswrapper[4931]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 29 00:18:06 crc kubenswrapper[4931]: > Nov 29 00:18:07 crc kubenswrapper[4931]: E1129 00:18:07.771204 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Nov 29 00:18:07 crc kubenswrapper[4931]: E1129 00:18:07.772344 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-825k8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-dj2pb_openstack(2c957f2c-45cc-4a11-8019-e20de202266b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 00:18:07 crc kubenswrapper[4931]: E1129 00:18:07.773596 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-dj2pb" podUID="2c957f2c-45cc-4a11-8019-e20de202266b" Nov 29 00:18:08 crc kubenswrapper[4931]: I1129 00:18:08.127943 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-trsph-config-v55dj"] Nov 29 00:18:08 crc kubenswrapper[4931]: I1129 00:18:08.698244 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0d661705-7bb5-482f-b2a8-79b44f6351ac","Type":"ContainerStarted","Data":"e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77"} Nov 29 00:18:08 crc kubenswrapper[4931]: I1129 00:18:08.698674 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 29 00:18:08 crc kubenswrapper[4931]: I1129 00:18:08.699983 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fbd78afc-2d6d-4118-92a2-e3499560359f","Type":"ContainerStarted","Data":"39fb76b6f4639dac5c87ee61fb096b14dda68bf6dce825287bbabae1121c3285"} Nov 29 00:18:08 crc kubenswrapper[4931]: I1129 00:18:08.700091 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:18:08 crc kubenswrapper[4931]: I1129 00:18:08.701727 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-trsph-config-v55dj" event={"ID":"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb","Type":"ContainerStarted","Data":"97a836d59e02d3efb3146f155216201d667c13ac2481ba56d3133411b26b3ae9"} Nov 29 00:18:08 crc kubenswrapper[4931]: I1129 00:18:08.701788 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-trsph-config-v55dj" event={"ID":"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb","Type":"ContainerStarted","Data":"59539c76ec0d585bcce327c439f5bf56946561c6aa88ea40189795f6b3b6e2eb"} Nov 29 00:18:08 crc kubenswrapper[4931]: E1129 00:18:08.703068 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-dj2pb" podUID="2c957f2c-45cc-4a11-8019-e20de202266b" Nov 29 00:18:08 crc kubenswrapper[4931]: I1129 00:18:08.724871 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=60.540090246 podStartE2EDuration="1m12.72485351s" podCreationTimestamp="2025-11-29 00:16:56 +0000 UTC" firstStartedPulling="2025-11-29 00:17:13.427671086 +0000 UTC m=+1046.589564318" lastFinishedPulling="2025-11-29 00:17:25.61243434 +0000 UTC m=+1058.774327582" observedRunningTime="2025-11-29 00:18:08.718146144 +0000 UTC m=+1101.880039396" watchObservedRunningTime="2025-11-29 00:18:08.72485351 +0000 UTC m=+1101.886746742" Nov 29 00:18:08 crc kubenswrapper[4931]: I1129 00:18:08.735657 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-trsph-config-v55dj" podStartSLOduration=6.735637129 podStartE2EDuration="6.735637129s" podCreationTimestamp="2025-11-29 00:18:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:18:08.733097809 +0000 UTC m=+1101.894991041" watchObservedRunningTime="2025-11-29 00:18:08.735637129 +0000 UTC m=+1101.897530371" Nov 29 00:18:08 crc kubenswrapper[4931]: I1129 00:18:08.766703 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=61.235914553 podStartE2EDuration="1m12.766685689s" podCreationTimestamp="2025-11-29 00:16:56 +0000 UTC" firstStartedPulling="2025-11-29 00:17:13.697471751 +0000 UTC m=+1046.859364993" lastFinishedPulling="2025-11-29 00:17:25.228242887 +0000 UTC m=+1058.390136129" observedRunningTime="2025-11-29 00:18:08.757019911 +0000 UTC m=+1101.918913163" watchObservedRunningTime="2025-11-29 00:18:08.766685689 +0000 UTC m=+1101.928578921" Nov 29 00:18:09 crc kubenswrapper[4931]: I1129 00:18:09.712379 4931 generic.go:334] "Generic (PLEG): container finished" podID="198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb" containerID="97a836d59e02d3efb3146f155216201d667c13ac2481ba56d3133411b26b3ae9" exitCode=0 Nov 29 00:18:09 crc kubenswrapper[4931]: I1129 00:18:09.712755 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-trsph-config-v55dj" event={"ID":"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb","Type":"ContainerDied","Data":"97a836d59e02d3efb3146f155216201d667c13ac2481ba56d3133411b26b3ae9"} Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.022225 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.186781 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-additional-scripts\") pod \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.186945 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-scripts\") pod \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.187012 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbpdf\" (UniqueName: \"kubernetes.io/projected/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-kube-api-access-xbpdf\") pod \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.187078 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-run-ovn\") pod \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.187252 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-log-ovn\") pod \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.187317 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-run\") pod \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\" (UID: \"198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb\") " Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.188069 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-run" (OuterVolumeSpecName: "var-run") pod "198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb" (UID: "198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.189353 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb" (UID: "198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.189458 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb" (UID: "198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.189508 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb" (UID: "198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.189738 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-scripts" (OuterVolumeSpecName: "scripts") pod "198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb" (UID: "198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.198554 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-kube-api-access-xbpdf" (OuterVolumeSpecName: "kube-api-access-xbpdf") pod "198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb" (UID: "198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb"). InnerVolumeSpecName "kube-api-access-xbpdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.255954 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-trsph-config-v55dj"] Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.265876 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-trsph-config-v55dj"] Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.290138 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.290211 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbpdf\" (UniqueName: \"kubernetes.io/projected/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-kube-api-access-xbpdf\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.290233 4931 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.291140 4931 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.291155 4931 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-var-run\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.291166 4931 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.727614 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59539c76ec0d585bcce327c439f5bf56946561c6aa88ea40189795f6b3b6e2eb" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.727924 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-trsph-config-v55dj" Nov 29 00:18:11 crc kubenswrapper[4931]: I1129 00:18:11.943148 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-trsph" Nov 29 00:18:13 crc kubenswrapper[4931]: I1129 00:18:13.231652 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb" path="/var/lib/kubelet/pods/198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb/volumes" Nov 29 00:18:15 crc kubenswrapper[4931]: I1129 00:18:15.863538 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:18:15 crc kubenswrapper[4931]: E1129 00:18:15.863793 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:18:15 crc kubenswrapper[4931]: E1129 00:18:15.863868 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 29 00:18:15 crc kubenswrapper[4931]: E1129 00:18:15.863957 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift podName:a720d879-8b09-452d-8371-9bf75f28cbde nodeName:}" failed. No retries permitted until 2025-11-29 00:18:47.863927404 +0000 UTC m=+1141.025820676 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift") pod "swift-storage-0" (UID: "a720d879-8b09-452d-8371-9bf75f28cbde") : configmap "swift-ring-files" not found Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.403184 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.652039 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.771498 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-kvkrd"] Nov 29 00:18:17 crc kubenswrapper[4931]: E1129 00:18:17.771914 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb" containerName="ovn-config" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.771931 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb" containerName="ovn-config" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.772139 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="198703a6-72c6-4fa6-b44b-0d6a1d4b5bbb" containerName="ovn-config" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.772748 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kvkrd" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.785502 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-kvkrd"] Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.862431 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-rrpld"] Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.863395 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rrpld" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.869636 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-86a9-account-create-update-n4ljg"] Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.870690 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-86a9-account-create-update-n4ljg" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.872991 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.894475 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-86a9-account-create-update-n4ljg"] Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.897394 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hrhq\" (UniqueName: \"kubernetes.io/projected/abb709fd-4b15-4b57-bebd-2e1f945c4334-kube-api-access-9hrhq\") pod \"cinder-db-create-kvkrd\" (UID: \"abb709fd-4b15-4b57-bebd-2e1f945c4334\") " pod="openstack/cinder-db-create-kvkrd" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.899845 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abb709fd-4b15-4b57-bebd-2e1f945c4334-operator-scripts\") pod \"cinder-db-create-kvkrd\" (UID: \"abb709fd-4b15-4b57-bebd-2e1f945c4334\") " pod="openstack/cinder-db-create-kvkrd" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.922333 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rrpld"] Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.958749 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-0c9c-account-create-update-6bwg6"] Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.963049 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0c9c-account-create-update-6bwg6" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.965944 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 29 00:18:17 crc kubenswrapper[4931]: I1129 00:18:17.970942 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0c9c-account-create-update-6bwg6"] Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.001601 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc9l8\" (UniqueName: \"kubernetes.io/projected/2d93d624-611c-405b-925c-41a3e159183d-kube-api-access-nc9l8\") pod \"barbican-db-create-rrpld\" (UID: \"2d93d624-611c-405b-925c-41a3e159183d\") " pod="openstack/barbican-db-create-rrpld" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.001833 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d93d624-611c-405b-925c-41a3e159183d-operator-scripts\") pod \"barbican-db-create-rrpld\" (UID: \"2d93d624-611c-405b-925c-41a3e159183d\") " pod="openstack/barbican-db-create-rrpld" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.002009 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3a60c30-b568-4927-bbe0-fcebead290cb-operator-scripts\") pod \"cinder-86a9-account-create-update-n4ljg\" (UID: \"e3a60c30-b568-4927-bbe0-fcebead290cb\") " pod="openstack/cinder-86a9-account-create-update-n4ljg" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.002104 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hrhq\" (UniqueName: \"kubernetes.io/projected/abb709fd-4b15-4b57-bebd-2e1f945c4334-kube-api-access-9hrhq\") pod \"cinder-db-create-kvkrd\" (UID: \"abb709fd-4b15-4b57-bebd-2e1f945c4334\") " pod="openstack/cinder-db-create-kvkrd" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.002204 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abb709fd-4b15-4b57-bebd-2e1f945c4334-operator-scripts\") pod \"cinder-db-create-kvkrd\" (UID: \"abb709fd-4b15-4b57-bebd-2e1f945c4334\") " pod="openstack/cinder-db-create-kvkrd" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.002279 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5glfd\" (UniqueName: \"kubernetes.io/projected/e3a60c30-b568-4927-bbe0-fcebead290cb-kube-api-access-5glfd\") pod \"cinder-86a9-account-create-update-n4ljg\" (UID: \"e3a60c30-b568-4927-bbe0-fcebead290cb\") " pod="openstack/cinder-86a9-account-create-update-n4ljg" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.002923 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abb709fd-4b15-4b57-bebd-2e1f945c4334-operator-scripts\") pod \"cinder-db-create-kvkrd\" (UID: \"abb709fd-4b15-4b57-bebd-2e1f945c4334\") " pod="openstack/cinder-db-create-kvkrd" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.028015 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hrhq\" (UniqueName: \"kubernetes.io/projected/abb709fd-4b15-4b57-bebd-2e1f945c4334-kube-api-access-9hrhq\") pod \"cinder-db-create-kvkrd\" (UID: \"abb709fd-4b15-4b57-bebd-2e1f945c4334\") " pod="openstack/cinder-db-create-kvkrd" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.069688 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-d4wgk"] Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.071007 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-d4wgk" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.081710 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-d4wgk"] Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.088537 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kvkrd" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.107646 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5glfd\" (UniqueName: \"kubernetes.io/projected/e3a60c30-b568-4927-bbe0-fcebead290cb-kube-api-access-5glfd\") pod \"cinder-86a9-account-create-update-n4ljg\" (UID: \"e3a60c30-b568-4927-bbe0-fcebead290cb\") " pod="openstack/cinder-86a9-account-create-update-n4ljg" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.107705 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc9l8\" (UniqueName: \"kubernetes.io/projected/2d93d624-611c-405b-925c-41a3e159183d-kube-api-access-nc9l8\") pod \"barbican-db-create-rrpld\" (UID: \"2d93d624-611c-405b-925c-41a3e159183d\") " pod="openstack/barbican-db-create-rrpld" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.107735 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d93d624-611c-405b-925c-41a3e159183d-operator-scripts\") pod \"barbican-db-create-rrpld\" (UID: \"2d93d624-611c-405b-925c-41a3e159183d\") " pod="openstack/barbican-db-create-rrpld" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.107863 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3a60c30-b568-4927-bbe0-fcebead290cb-operator-scripts\") pod \"cinder-86a9-account-create-update-n4ljg\" (UID: \"e3a60c30-b568-4927-bbe0-fcebead290cb\") " pod="openstack/cinder-86a9-account-create-update-n4ljg" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.107887 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xhfd\" (UniqueName: \"kubernetes.io/projected/92aad38f-3527-480c-ac71-23043a012568-kube-api-access-7xhfd\") pod \"barbican-0c9c-account-create-update-6bwg6\" (UID: \"92aad38f-3527-480c-ac71-23043a012568\") " pod="openstack/barbican-0c9c-account-create-update-6bwg6" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.107923 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92aad38f-3527-480c-ac71-23043a012568-operator-scripts\") pod \"barbican-0c9c-account-create-update-6bwg6\" (UID: \"92aad38f-3527-480c-ac71-23043a012568\") " pod="openstack/barbican-0c9c-account-create-update-6bwg6" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.108788 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d93d624-611c-405b-925c-41a3e159183d-operator-scripts\") pod \"barbican-db-create-rrpld\" (UID: \"2d93d624-611c-405b-925c-41a3e159183d\") " pod="openstack/barbican-db-create-rrpld" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.108910 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3a60c30-b568-4927-bbe0-fcebead290cb-operator-scripts\") pod \"cinder-86a9-account-create-update-n4ljg\" (UID: \"e3a60c30-b568-4927-bbe0-fcebead290cb\") " pod="openstack/cinder-86a9-account-create-update-n4ljg" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.133622 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-ltd4m"] Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.134840 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.141591 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bwpgd" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.142022 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.143036 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.143088 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.148423 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc9l8\" (UniqueName: \"kubernetes.io/projected/2d93d624-611c-405b-925c-41a3e159183d-kube-api-access-nc9l8\") pod \"barbican-db-create-rrpld\" (UID: \"2d93d624-611c-405b-925c-41a3e159183d\") " pod="openstack/barbican-db-create-rrpld" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.149911 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5glfd\" (UniqueName: \"kubernetes.io/projected/e3a60c30-b568-4927-bbe0-fcebead290cb-kube-api-access-5glfd\") pod \"cinder-86a9-account-create-update-n4ljg\" (UID: \"e3a60c30-b568-4927-bbe0-fcebead290cb\") " pod="openstack/cinder-86a9-account-create-update-n4ljg" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.157518 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ltd4m"] Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.190881 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rrpld" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.194355 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-86a9-account-create-update-n4ljg" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.209521 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/897f5cd4-7cba-452f-bf48-ebe6c3333b76-operator-scripts\") pod \"neutron-db-create-d4wgk\" (UID: \"897f5cd4-7cba-452f-bf48-ebe6c3333b76\") " pod="openstack/neutron-db-create-d4wgk" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.209605 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvlb9\" (UniqueName: \"kubernetes.io/projected/fae16d0f-aa7b-4f9b-aff2-896708ff5061-kube-api-access-lvlb9\") pod \"keystone-db-sync-ltd4m\" (UID: \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\") " pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.209746 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xhfd\" (UniqueName: \"kubernetes.io/projected/92aad38f-3527-480c-ac71-23043a012568-kube-api-access-7xhfd\") pod \"barbican-0c9c-account-create-update-6bwg6\" (UID: \"92aad38f-3527-480c-ac71-23043a012568\") " pod="openstack/barbican-0c9c-account-create-update-6bwg6" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.209787 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae16d0f-aa7b-4f9b-aff2-896708ff5061-combined-ca-bundle\") pod \"keystone-db-sync-ltd4m\" (UID: \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\") " pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.209874 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92aad38f-3527-480c-ac71-23043a012568-operator-scripts\") pod \"barbican-0c9c-account-create-update-6bwg6\" (UID: \"92aad38f-3527-480c-ac71-23043a012568\") " pod="openstack/barbican-0c9c-account-create-update-6bwg6" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.209966 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h57sc\" (UniqueName: \"kubernetes.io/projected/897f5cd4-7cba-452f-bf48-ebe6c3333b76-kube-api-access-h57sc\") pod \"neutron-db-create-d4wgk\" (UID: \"897f5cd4-7cba-452f-bf48-ebe6c3333b76\") " pod="openstack/neutron-db-create-d4wgk" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.209993 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae16d0f-aa7b-4f9b-aff2-896708ff5061-config-data\") pod \"keystone-db-sync-ltd4m\" (UID: \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\") " pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.211363 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92aad38f-3527-480c-ac71-23043a012568-operator-scripts\") pod \"barbican-0c9c-account-create-update-6bwg6\" (UID: \"92aad38f-3527-480c-ac71-23043a012568\") " pod="openstack/barbican-0c9c-account-create-update-6bwg6" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.239442 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xhfd\" (UniqueName: \"kubernetes.io/projected/92aad38f-3527-480c-ac71-23043a012568-kube-api-access-7xhfd\") pod \"barbican-0c9c-account-create-update-6bwg6\" (UID: \"92aad38f-3527-480c-ac71-23043a012568\") " pod="openstack/barbican-0c9c-account-create-update-6bwg6" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.270031 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-2d72-account-create-update-qb5cv"] Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.271348 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2d72-account-create-update-qb5cv" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.275022 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.282332 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0c9c-account-create-update-6bwg6" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.287578 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2d72-account-create-update-qb5cv"] Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.323569 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae16d0f-aa7b-4f9b-aff2-896708ff5061-combined-ca-bundle\") pod \"keystone-db-sync-ltd4m\" (UID: \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\") " pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.323672 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h57sc\" (UniqueName: \"kubernetes.io/projected/897f5cd4-7cba-452f-bf48-ebe6c3333b76-kube-api-access-h57sc\") pod \"neutron-db-create-d4wgk\" (UID: \"897f5cd4-7cba-452f-bf48-ebe6c3333b76\") " pod="openstack/neutron-db-create-d4wgk" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.323692 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae16d0f-aa7b-4f9b-aff2-896708ff5061-config-data\") pod \"keystone-db-sync-ltd4m\" (UID: \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\") " pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.323769 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/897f5cd4-7cba-452f-bf48-ebe6c3333b76-operator-scripts\") pod \"neutron-db-create-d4wgk\" (UID: \"897f5cd4-7cba-452f-bf48-ebe6c3333b76\") " pod="openstack/neutron-db-create-d4wgk" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.323824 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvlb9\" (UniqueName: \"kubernetes.io/projected/fae16d0f-aa7b-4f9b-aff2-896708ff5061-kube-api-access-lvlb9\") pod \"keystone-db-sync-ltd4m\" (UID: \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\") " pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.325195 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/897f5cd4-7cba-452f-bf48-ebe6c3333b76-operator-scripts\") pod \"neutron-db-create-d4wgk\" (UID: \"897f5cd4-7cba-452f-bf48-ebe6c3333b76\") " pod="openstack/neutron-db-create-d4wgk" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.327595 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae16d0f-aa7b-4f9b-aff2-896708ff5061-combined-ca-bundle\") pod \"keystone-db-sync-ltd4m\" (UID: \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\") " pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.330533 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae16d0f-aa7b-4f9b-aff2-896708ff5061-config-data\") pod \"keystone-db-sync-ltd4m\" (UID: \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\") " pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.344305 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvlb9\" (UniqueName: \"kubernetes.io/projected/fae16d0f-aa7b-4f9b-aff2-896708ff5061-kube-api-access-lvlb9\") pod \"keystone-db-sync-ltd4m\" (UID: \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\") " pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.359316 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h57sc\" (UniqueName: \"kubernetes.io/projected/897f5cd4-7cba-452f-bf48-ebe6c3333b76-kube-api-access-h57sc\") pod \"neutron-db-create-d4wgk\" (UID: \"897f5cd4-7cba-452f-bf48-ebe6c3333b76\") " pod="openstack/neutron-db-create-d4wgk" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.395201 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-d4wgk" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.427824 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d367827-344f-49b6-9b67-9e7b08d2c89f-operator-scripts\") pod \"neutron-2d72-account-create-update-qb5cv\" (UID: \"1d367827-344f-49b6-9b67-9e7b08d2c89f\") " pod="openstack/neutron-2d72-account-create-update-qb5cv" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.427887 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlfh9\" (UniqueName: \"kubernetes.io/projected/1d367827-344f-49b6-9b67-9e7b08d2c89f-kube-api-access-nlfh9\") pod \"neutron-2d72-account-create-update-qb5cv\" (UID: \"1d367827-344f-49b6-9b67-9e7b08d2c89f\") " pod="openstack/neutron-2d72-account-create-update-qb5cv" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.528964 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d367827-344f-49b6-9b67-9e7b08d2c89f-operator-scripts\") pod \"neutron-2d72-account-create-update-qb5cv\" (UID: \"1d367827-344f-49b6-9b67-9e7b08d2c89f\") " pod="openstack/neutron-2d72-account-create-update-qb5cv" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.529034 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlfh9\" (UniqueName: \"kubernetes.io/projected/1d367827-344f-49b6-9b67-9e7b08d2c89f-kube-api-access-nlfh9\") pod \"neutron-2d72-account-create-update-qb5cv\" (UID: \"1d367827-344f-49b6-9b67-9e7b08d2c89f\") " pod="openstack/neutron-2d72-account-create-update-qb5cv" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.529582 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d367827-344f-49b6-9b67-9e7b08d2c89f-operator-scripts\") pod \"neutron-2d72-account-create-update-qb5cv\" (UID: \"1d367827-344f-49b6-9b67-9e7b08d2c89f\") " pod="openstack/neutron-2d72-account-create-update-qb5cv" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.548989 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlfh9\" (UniqueName: \"kubernetes.io/projected/1d367827-344f-49b6-9b67-9e7b08d2c89f-kube-api-access-nlfh9\") pod \"neutron-2d72-account-create-update-qb5cv\" (UID: \"1d367827-344f-49b6-9b67-9e7b08d2c89f\") " pod="openstack/neutron-2d72-account-create-update-qb5cv" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.593761 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.603915 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2d72-account-create-update-qb5cv" Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.651381 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-kvkrd"] Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.747751 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rrpld"] Nov 29 00:18:18 crc kubenswrapper[4931]: W1129 00:18:18.765403 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d93d624_611c_405b_925c_41a3e159183d.slice/crio-f2968ba67cc0151733dbccfea4cb70d883516de6f1af7162934e53e2262b7a92 WatchSource:0}: Error finding container f2968ba67cc0151733dbccfea4cb70d883516de6f1af7162934e53e2262b7a92: Status 404 returned error can't find the container with id f2968ba67cc0151733dbccfea4cb70d883516de6f1af7162934e53e2262b7a92 Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.799703 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rrpld" event={"ID":"2d93d624-611c-405b-925c-41a3e159183d","Type":"ContainerStarted","Data":"f2968ba67cc0151733dbccfea4cb70d883516de6f1af7162934e53e2262b7a92"} Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.801312 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-kvkrd" event={"ID":"abb709fd-4b15-4b57-bebd-2e1f945c4334","Type":"ContainerStarted","Data":"d35aa87a329d60df26625996a08fb8d3bea5d7edea95264bdbdce21e50113bac"} Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.838432 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-86a9-account-create-update-n4ljg"] Nov 29 00:18:18 crc kubenswrapper[4931]: I1129 00:18:18.851704 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0c9c-account-create-update-6bwg6"] Nov 29 00:18:18 crc kubenswrapper[4931]: W1129 00:18:18.858992 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92aad38f_3527_480c_ac71_23043a012568.slice/crio-f70631fa1fe04edbd3cced9e036bfd5b04d8e9ff01f7d3d1fcf7a5bcf843ef59 WatchSource:0}: Error finding container f70631fa1fe04edbd3cced9e036bfd5b04d8e9ff01f7d3d1fcf7a5bcf843ef59: Status 404 returned error can't find the container with id f70631fa1fe04edbd3cced9e036bfd5b04d8e9ff01f7d3d1fcf7a5bcf843ef59 Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.001078 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-d4wgk"] Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.092558 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ltd4m"] Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.115761 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2d72-account-create-update-qb5cv"] Nov 29 00:18:19 crc kubenswrapper[4931]: W1129 00:18:19.125154 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d367827_344f_49b6_9b67_9e7b08d2c89f.slice/crio-72634ab0043c7939e8b73a9deb92f0d60a78fa0c3ea4d1d8a9effa7486e9be4f WatchSource:0}: Error finding container 72634ab0043c7939e8b73a9deb92f0d60a78fa0c3ea4d1d8a9effa7486e9be4f: Status 404 returned error can't find the container with id 72634ab0043c7939e8b73a9deb92f0d60a78fa0c3ea4d1d8a9effa7486e9be4f Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.810271 4931 generic.go:334] "Generic (PLEG): container finished" podID="e3a60c30-b568-4927-bbe0-fcebead290cb" containerID="23bcf002d1c5753522b14b9cebb06be0abf86762a0958c6bc9a4a018808cd16b" exitCode=0 Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.810457 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-86a9-account-create-update-n4ljg" event={"ID":"e3a60c30-b568-4927-bbe0-fcebead290cb","Type":"ContainerDied","Data":"23bcf002d1c5753522b14b9cebb06be0abf86762a0958c6bc9a4a018808cd16b"} Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.810616 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-86a9-account-create-update-n4ljg" event={"ID":"e3a60c30-b568-4927-bbe0-fcebead290cb","Type":"ContainerStarted","Data":"1f646aa33d76674eade1c98404da5f0e7d32443354d3562cac44d3e4aa149164"} Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.812369 4931 generic.go:334] "Generic (PLEG): container finished" podID="897f5cd4-7cba-452f-bf48-ebe6c3333b76" containerID="d25680c15e3bf4b84c3c75307bee7854603df02a9c964738af8672da521e16e3" exitCode=0 Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.812464 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-d4wgk" event={"ID":"897f5cd4-7cba-452f-bf48-ebe6c3333b76","Type":"ContainerDied","Data":"d25680c15e3bf4b84c3c75307bee7854603df02a9c964738af8672da521e16e3"} Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.812521 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-d4wgk" event={"ID":"897f5cd4-7cba-452f-bf48-ebe6c3333b76","Type":"ContainerStarted","Data":"63623563df96400e6d98a0bf0b687895eb95deece2a7a81540b03c5013555598"} Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.814436 4931 generic.go:334] "Generic (PLEG): container finished" podID="92aad38f-3527-480c-ac71-23043a012568" containerID="46a7a92d8f936282a47882ed62369636caac887e41020f2f68642a95cd156509" exitCode=0 Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.814505 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0c9c-account-create-update-6bwg6" event={"ID":"92aad38f-3527-480c-ac71-23043a012568","Type":"ContainerDied","Data":"46a7a92d8f936282a47882ed62369636caac887e41020f2f68642a95cd156509"} Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.814534 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0c9c-account-create-update-6bwg6" event={"ID":"92aad38f-3527-480c-ac71-23043a012568","Type":"ContainerStarted","Data":"f70631fa1fe04edbd3cced9e036bfd5b04d8e9ff01f7d3d1fcf7a5bcf843ef59"} Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.816147 4931 generic.go:334] "Generic (PLEG): container finished" podID="2d93d624-611c-405b-925c-41a3e159183d" containerID="e0fd7cff7473147e00ffdc10c155c7d1bc55e6e49e7189c3ef409d989ab1d86a" exitCode=0 Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.816186 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rrpld" event={"ID":"2d93d624-611c-405b-925c-41a3e159183d","Type":"ContainerDied","Data":"e0fd7cff7473147e00ffdc10c155c7d1bc55e6e49e7189c3ef409d989ab1d86a"} Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.817523 4931 generic.go:334] "Generic (PLEG): container finished" podID="1d367827-344f-49b6-9b67-9e7b08d2c89f" containerID="0c5009dbbe024a398e932da23bd1e88284cd87ee64660cab3a23aacacd9f7db3" exitCode=0 Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.817582 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2d72-account-create-update-qb5cv" event={"ID":"1d367827-344f-49b6-9b67-9e7b08d2c89f","Type":"ContainerDied","Data":"0c5009dbbe024a398e932da23bd1e88284cd87ee64660cab3a23aacacd9f7db3"} Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.817612 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2d72-account-create-update-qb5cv" event={"ID":"1d367827-344f-49b6-9b67-9e7b08d2c89f","Type":"ContainerStarted","Data":"72634ab0043c7939e8b73a9deb92f0d60a78fa0c3ea4d1d8a9effa7486e9be4f"} Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.818684 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ltd4m" event={"ID":"fae16d0f-aa7b-4f9b-aff2-896708ff5061","Type":"ContainerStarted","Data":"6318464376fb7f5f51f793e785e0545034b307c19071dd86a257624eced995ff"} Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.820032 4931 generic.go:334] "Generic (PLEG): container finished" podID="abb709fd-4b15-4b57-bebd-2e1f945c4334" containerID="6f88355e1883ba361cc10dfbeeb09a2bce98d2f2d3f6e2d4108ff6798c15607e" exitCode=0 Nov 29 00:18:19 crc kubenswrapper[4931]: I1129 00:18:19.820064 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-kvkrd" event={"ID":"abb709fd-4b15-4b57-bebd-2e1f945c4334","Type":"ContainerDied","Data":"6f88355e1883ba361cc10dfbeeb09a2bce98d2f2d3f6e2d4108ff6798c15607e"} Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.243383 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0c9c-account-create-update-6bwg6" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.438214 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xhfd\" (UniqueName: \"kubernetes.io/projected/92aad38f-3527-480c-ac71-23043a012568-kube-api-access-7xhfd\") pod \"92aad38f-3527-480c-ac71-23043a012568\" (UID: \"92aad38f-3527-480c-ac71-23043a012568\") " Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.438326 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92aad38f-3527-480c-ac71-23043a012568-operator-scripts\") pod \"92aad38f-3527-480c-ac71-23043a012568\" (UID: \"92aad38f-3527-480c-ac71-23043a012568\") " Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.439388 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92aad38f-3527-480c-ac71-23043a012568-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "92aad38f-3527-480c-ac71-23043a012568" (UID: "92aad38f-3527-480c-ac71-23043a012568"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.445132 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92aad38f-3527-480c-ac71-23043a012568-kube-api-access-7xhfd" (OuterVolumeSpecName: "kube-api-access-7xhfd") pod "92aad38f-3527-480c-ac71-23043a012568" (UID: "92aad38f-3527-480c-ac71-23043a012568"). InnerVolumeSpecName "kube-api-access-7xhfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.445406 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-d4wgk" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.449331 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-86a9-account-create-update-n4ljg" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.469592 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2d72-account-create-update-qb5cv" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.488225 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rrpld" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.494631 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kvkrd" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.540493 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xhfd\" (UniqueName: \"kubernetes.io/projected/92aad38f-3527-480c-ac71-23043a012568-kube-api-access-7xhfd\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.540537 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92aad38f-3527-480c-ac71-23043a012568-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.641344 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/897f5cd4-7cba-452f-bf48-ebe6c3333b76-operator-scripts\") pod \"897f5cd4-7cba-452f-bf48-ebe6c3333b76\" (UID: \"897f5cd4-7cba-452f-bf48-ebe6c3333b76\") " Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.641484 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abb709fd-4b15-4b57-bebd-2e1f945c4334-operator-scripts\") pod \"abb709fd-4b15-4b57-bebd-2e1f945c4334\" (UID: \"abb709fd-4b15-4b57-bebd-2e1f945c4334\") " Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.641528 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5glfd\" (UniqueName: \"kubernetes.io/projected/e3a60c30-b568-4927-bbe0-fcebead290cb-kube-api-access-5glfd\") pod \"e3a60c30-b568-4927-bbe0-fcebead290cb\" (UID: \"e3a60c30-b568-4927-bbe0-fcebead290cb\") " Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.641573 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3a60c30-b568-4927-bbe0-fcebead290cb-operator-scripts\") pod \"e3a60c30-b568-4927-bbe0-fcebead290cb\" (UID: \"e3a60c30-b568-4927-bbe0-fcebead290cb\") " Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.641627 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d93d624-611c-405b-925c-41a3e159183d-operator-scripts\") pod \"2d93d624-611c-405b-925c-41a3e159183d\" (UID: \"2d93d624-611c-405b-925c-41a3e159183d\") " Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.641650 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hrhq\" (UniqueName: \"kubernetes.io/projected/abb709fd-4b15-4b57-bebd-2e1f945c4334-kube-api-access-9hrhq\") pod \"abb709fd-4b15-4b57-bebd-2e1f945c4334\" (UID: \"abb709fd-4b15-4b57-bebd-2e1f945c4334\") " Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.641702 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d367827-344f-49b6-9b67-9e7b08d2c89f-operator-scripts\") pod \"1d367827-344f-49b6-9b67-9e7b08d2c89f\" (UID: \"1d367827-344f-49b6-9b67-9e7b08d2c89f\") " Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.641733 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlfh9\" (UniqueName: \"kubernetes.io/projected/1d367827-344f-49b6-9b67-9e7b08d2c89f-kube-api-access-nlfh9\") pod \"1d367827-344f-49b6-9b67-9e7b08d2c89f\" (UID: \"1d367827-344f-49b6-9b67-9e7b08d2c89f\") " Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.641797 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc9l8\" (UniqueName: \"kubernetes.io/projected/2d93d624-611c-405b-925c-41a3e159183d-kube-api-access-nc9l8\") pod \"2d93d624-611c-405b-925c-41a3e159183d\" (UID: \"2d93d624-611c-405b-925c-41a3e159183d\") " Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.642043 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h57sc\" (UniqueName: \"kubernetes.io/projected/897f5cd4-7cba-452f-bf48-ebe6c3333b76-kube-api-access-h57sc\") pod \"897f5cd4-7cba-452f-bf48-ebe6c3333b76\" (UID: \"897f5cd4-7cba-452f-bf48-ebe6c3333b76\") " Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.642887 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/897f5cd4-7cba-452f-bf48-ebe6c3333b76-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "897f5cd4-7cba-452f-bf48-ebe6c3333b76" (UID: "897f5cd4-7cba-452f-bf48-ebe6c3333b76"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.642946 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3a60c30-b568-4927-bbe0-fcebead290cb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e3a60c30-b568-4927-bbe0-fcebead290cb" (UID: "e3a60c30-b568-4927-bbe0-fcebead290cb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.645128 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d367827-344f-49b6-9b67-9e7b08d2c89f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1d367827-344f-49b6-9b67-9e7b08d2c89f" (UID: "1d367827-344f-49b6-9b67-9e7b08d2c89f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.645165 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d93d624-611c-405b-925c-41a3e159183d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d93d624-611c-405b-925c-41a3e159183d" (UID: "2d93d624-611c-405b-925c-41a3e159183d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.645790 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3a60c30-b568-4927-bbe0-fcebead290cb-kube-api-access-5glfd" (OuterVolumeSpecName: "kube-api-access-5glfd") pod "e3a60c30-b568-4927-bbe0-fcebead290cb" (UID: "e3a60c30-b568-4927-bbe0-fcebead290cb"). InnerVolumeSpecName "kube-api-access-5glfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.645942 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d93d624-611c-405b-925c-41a3e159183d-kube-api-access-nc9l8" (OuterVolumeSpecName: "kube-api-access-nc9l8") pod "2d93d624-611c-405b-925c-41a3e159183d" (UID: "2d93d624-611c-405b-925c-41a3e159183d"). InnerVolumeSpecName "kube-api-access-nc9l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.648072 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d367827-344f-49b6-9b67-9e7b08d2c89f-kube-api-access-nlfh9" (OuterVolumeSpecName: "kube-api-access-nlfh9") pod "1d367827-344f-49b6-9b67-9e7b08d2c89f" (UID: "1d367827-344f-49b6-9b67-9e7b08d2c89f"). InnerVolumeSpecName "kube-api-access-nlfh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.648131 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abb709fd-4b15-4b57-bebd-2e1f945c4334-kube-api-access-9hrhq" (OuterVolumeSpecName: "kube-api-access-9hrhq") pod "abb709fd-4b15-4b57-bebd-2e1f945c4334" (UID: "abb709fd-4b15-4b57-bebd-2e1f945c4334"). InnerVolumeSpecName "kube-api-access-9hrhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.653329 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abb709fd-4b15-4b57-bebd-2e1f945c4334-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "abb709fd-4b15-4b57-bebd-2e1f945c4334" (UID: "abb709fd-4b15-4b57-bebd-2e1f945c4334"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.656282 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/897f5cd4-7cba-452f-bf48-ebe6c3333b76-kube-api-access-h57sc" (OuterVolumeSpecName: "kube-api-access-h57sc") pod "897f5cd4-7cba-452f-bf48-ebe6c3333b76" (UID: "897f5cd4-7cba-452f-bf48-ebe6c3333b76"). InnerVolumeSpecName "kube-api-access-h57sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.745261 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlfh9\" (UniqueName: \"kubernetes.io/projected/1d367827-344f-49b6-9b67-9e7b08d2c89f-kube-api-access-nlfh9\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.745293 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc9l8\" (UniqueName: \"kubernetes.io/projected/2d93d624-611c-405b-925c-41a3e159183d-kube-api-access-nc9l8\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.745303 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h57sc\" (UniqueName: \"kubernetes.io/projected/897f5cd4-7cba-452f-bf48-ebe6c3333b76-kube-api-access-h57sc\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.745313 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/897f5cd4-7cba-452f-bf48-ebe6c3333b76-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.745322 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abb709fd-4b15-4b57-bebd-2e1f945c4334-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.745332 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5glfd\" (UniqueName: \"kubernetes.io/projected/e3a60c30-b568-4927-bbe0-fcebead290cb-kube-api-access-5glfd\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.745340 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e3a60c30-b568-4927-bbe0-fcebead290cb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.745348 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d93d624-611c-405b-925c-41a3e159183d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.745356 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hrhq\" (UniqueName: \"kubernetes.io/projected/abb709fd-4b15-4b57-bebd-2e1f945c4334-kube-api-access-9hrhq\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.745365 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d367827-344f-49b6-9b67-9e7b08d2c89f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.839560 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-86a9-account-create-update-n4ljg" event={"ID":"e3a60c30-b568-4927-bbe0-fcebead290cb","Type":"ContainerDied","Data":"1f646aa33d76674eade1c98404da5f0e7d32443354d3562cac44d3e4aa149164"} Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.839604 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f646aa33d76674eade1c98404da5f0e7d32443354d3562cac44d3e4aa149164" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.839603 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-86a9-account-create-update-n4ljg" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.841106 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-d4wgk" event={"ID":"897f5cd4-7cba-452f-bf48-ebe6c3333b76","Type":"ContainerDied","Data":"63623563df96400e6d98a0bf0b687895eb95deece2a7a81540b03c5013555598"} Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.841134 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63623563df96400e6d98a0bf0b687895eb95deece2a7a81540b03c5013555598" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.841181 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-d4wgk" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.879597 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0c9c-account-create-update-6bwg6" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.880126 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0c9c-account-create-update-6bwg6" event={"ID":"92aad38f-3527-480c-ac71-23043a012568","Type":"ContainerDied","Data":"f70631fa1fe04edbd3cced9e036bfd5b04d8e9ff01f7d3d1fcf7a5bcf843ef59"} Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.880277 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f70631fa1fe04edbd3cced9e036bfd5b04d8e9ff01f7d3d1fcf7a5bcf843ef59" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.889108 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rrpld" event={"ID":"2d93d624-611c-405b-925c-41a3e159183d","Type":"ContainerDied","Data":"f2968ba67cc0151733dbccfea4cb70d883516de6f1af7162934e53e2262b7a92"} Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.889164 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2968ba67cc0151733dbccfea4cb70d883516de6f1af7162934e53e2262b7a92" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.889236 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rrpld" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.904041 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2d72-account-create-update-qb5cv" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.904314 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2d72-account-create-update-qb5cv" event={"ID":"1d367827-344f-49b6-9b67-9e7b08d2c89f","Type":"ContainerDied","Data":"72634ab0043c7939e8b73a9deb92f0d60a78fa0c3ea4d1d8a9effa7486e9be4f"} Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.905455 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72634ab0043c7939e8b73a9deb92f0d60a78fa0c3ea4d1d8a9effa7486e9be4f" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.907969 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-kvkrd" event={"ID":"abb709fd-4b15-4b57-bebd-2e1f945c4334","Type":"ContainerDied","Data":"d35aa87a329d60df26625996a08fb8d3bea5d7edea95264bdbdce21e50113bac"} Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.908007 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d35aa87a329d60df26625996a08fb8d3bea5d7edea95264bdbdce21e50113bac" Nov 29 00:18:21 crc kubenswrapper[4931]: I1129 00:18:21.908061 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kvkrd" Nov 29 00:18:25 crc kubenswrapper[4931]: I1129 00:18:25.960705 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ltd4m" event={"ID":"fae16d0f-aa7b-4f9b-aff2-896708ff5061","Type":"ContainerStarted","Data":"01d7fd77664235a8f6c4e5799750b304ad66c6e43d7e9dda8d763f02f0a460c1"} Nov 29 00:18:25 crc kubenswrapper[4931]: I1129 00:18:25.962513 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-dj2pb" event={"ID":"2c957f2c-45cc-4a11-8019-e20de202266b","Type":"ContainerStarted","Data":"549aecf595ecff89023b1a2b77d9b714fa543e1a33c0e1a53ad670b5080a7092"} Nov 29 00:18:25 crc kubenswrapper[4931]: I1129 00:18:25.977088 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-ltd4m" podStartSLOduration=1.821034387 podStartE2EDuration="7.97707234s" podCreationTimestamp="2025-11-29 00:18:18 +0000 UTC" firstStartedPulling="2025-11-29 00:18:19.10857364 +0000 UTC m=+1112.270466872" lastFinishedPulling="2025-11-29 00:18:25.264611573 +0000 UTC m=+1118.426504825" observedRunningTime="2025-11-29 00:18:25.974440665 +0000 UTC m=+1119.136333897" watchObservedRunningTime="2025-11-29 00:18:25.97707234 +0000 UTC m=+1119.138965572" Nov 29 00:18:29 crc kubenswrapper[4931]: I1129 00:18:29.990168 4931 generic.go:334] "Generic (PLEG): container finished" podID="fae16d0f-aa7b-4f9b-aff2-896708ff5061" containerID="01d7fd77664235a8f6c4e5799750b304ad66c6e43d7e9dda8d763f02f0a460c1" exitCode=0 Nov 29 00:18:29 crc kubenswrapper[4931]: I1129 00:18:29.990258 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ltd4m" event={"ID":"fae16d0f-aa7b-4f9b-aff2-896708ff5061","Type":"ContainerDied","Data":"01d7fd77664235a8f6c4e5799750b304ad66c6e43d7e9dda8d763f02f0a460c1"} Nov 29 00:18:30 crc kubenswrapper[4931]: I1129 00:18:30.021800 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-dj2pb" podStartSLOduration=6.247226947 podStartE2EDuration="39.021767348s" podCreationTimestamp="2025-11-29 00:17:51 +0000 UTC" firstStartedPulling="2025-11-29 00:17:52.469827103 +0000 UTC m=+1085.631720355" lastFinishedPulling="2025-11-29 00:18:25.244367484 +0000 UTC m=+1118.406260756" observedRunningTime="2025-11-29 00:18:25.996248719 +0000 UTC m=+1119.158142011" watchObservedRunningTime="2025-11-29 00:18:30.021767348 +0000 UTC m=+1123.183660660" Nov 29 00:18:31 crc kubenswrapper[4931]: I1129 00:18:31.369617 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:31 crc kubenswrapper[4931]: I1129 00:18:31.456295 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvlb9\" (UniqueName: \"kubernetes.io/projected/fae16d0f-aa7b-4f9b-aff2-896708ff5061-kube-api-access-lvlb9\") pod \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\" (UID: \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\") " Nov 29 00:18:31 crc kubenswrapper[4931]: I1129 00:18:31.456370 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae16d0f-aa7b-4f9b-aff2-896708ff5061-combined-ca-bundle\") pod \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\" (UID: \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\") " Nov 29 00:18:31 crc kubenswrapper[4931]: I1129 00:18:31.456488 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae16d0f-aa7b-4f9b-aff2-896708ff5061-config-data\") pod \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\" (UID: \"fae16d0f-aa7b-4f9b-aff2-896708ff5061\") " Nov 29 00:18:31 crc kubenswrapper[4931]: I1129 00:18:31.465656 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fae16d0f-aa7b-4f9b-aff2-896708ff5061-kube-api-access-lvlb9" (OuterVolumeSpecName: "kube-api-access-lvlb9") pod "fae16d0f-aa7b-4f9b-aff2-896708ff5061" (UID: "fae16d0f-aa7b-4f9b-aff2-896708ff5061"). InnerVolumeSpecName "kube-api-access-lvlb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:18:31 crc kubenswrapper[4931]: I1129 00:18:31.486352 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fae16d0f-aa7b-4f9b-aff2-896708ff5061-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fae16d0f-aa7b-4f9b-aff2-896708ff5061" (UID: "fae16d0f-aa7b-4f9b-aff2-896708ff5061"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:18:31 crc kubenswrapper[4931]: I1129 00:18:31.511272 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fae16d0f-aa7b-4f9b-aff2-896708ff5061-config-data" (OuterVolumeSpecName: "config-data") pod "fae16d0f-aa7b-4f9b-aff2-896708ff5061" (UID: "fae16d0f-aa7b-4f9b-aff2-896708ff5061"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:18:31 crc kubenswrapper[4931]: I1129 00:18:31.560050 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvlb9\" (UniqueName: \"kubernetes.io/projected/fae16d0f-aa7b-4f9b-aff2-896708ff5061-kube-api-access-lvlb9\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:31 crc kubenswrapper[4931]: I1129 00:18:31.560314 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fae16d0f-aa7b-4f9b-aff2-896708ff5061-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:31 crc kubenswrapper[4931]: I1129 00:18:31.560501 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fae16d0f-aa7b-4f9b-aff2-896708ff5061-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.009890 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ltd4m" event={"ID":"fae16d0f-aa7b-4f9b-aff2-896708ff5061","Type":"ContainerDied","Data":"6318464376fb7f5f51f793e785e0545034b307c19071dd86a257624eced995ff"} Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.010909 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6318464376fb7f5f51f793e785e0545034b307c19071dd86a257624eced995ff" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.010596 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ltd4m" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.260536 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-r5sgw"] Nov 29 00:18:32 crc kubenswrapper[4931]: E1129 00:18:32.260838 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3a60c30-b568-4927-bbe0-fcebead290cb" containerName="mariadb-account-create-update" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.260849 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3a60c30-b568-4927-bbe0-fcebead290cb" containerName="mariadb-account-create-update" Nov 29 00:18:32 crc kubenswrapper[4931]: E1129 00:18:32.260855 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92aad38f-3527-480c-ac71-23043a012568" containerName="mariadb-account-create-update" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.260862 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="92aad38f-3527-480c-ac71-23043a012568" containerName="mariadb-account-create-update" Nov 29 00:18:32 crc kubenswrapper[4931]: E1129 00:18:32.260875 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d93d624-611c-405b-925c-41a3e159183d" containerName="mariadb-database-create" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.260881 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d93d624-611c-405b-925c-41a3e159183d" containerName="mariadb-database-create" Nov 29 00:18:32 crc kubenswrapper[4931]: E1129 00:18:32.260891 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb709fd-4b15-4b57-bebd-2e1f945c4334" containerName="mariadb-database-create" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.260897 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb709fd-4b15-4b57-bebd-2e1f945c4334" containerName="mariadb-database-create" Nov 29 00:18:32 crc kubenswrapper[4931]: E1129 00:18:32.260914 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fae16d0f-aa7b-4f9b-aff2-896708ff5061" containerName="keystone-db-sync" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.260920 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="fae16d0f-aa7b-4f9b-aff2-896708ff5061" containerName="keystone-db-sync" Nov 29 00:18:32 crc kubenswrapper[4931]: E1129 00:18:32.260929 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="897f5cd4-7cba-452f-bf48-ebe6c3333b76" containerName="mariadb-database-create" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.260935 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="897f5cd4-7cba-452f-bf48-ebe6c3333b76" containerName="mariadb-database-create" Nov 29 00:18:32 crc kubenswrapper[4931]: E1129 00:18:32.260943 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d367827-344f-49b6-9b67-9e7b08d2c89f" containerName="mariadb-account-create-update" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.260950 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d367827-344f-49b6-9b67-9e7b08d2c89f" containerName="mariadb-account-create-update" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.261112 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d93d624-611c-405b-925c-41a3e159183d" containerName="mariadb-database-create" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.261121 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="fae16d0f-aa7b-4f9b-aff2-896708ff5061" containerName="keystone-db-sync" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.261128 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="92aad38f-3527-480c-ac71-23043a012568" containerName="mariadb-account-create-update" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.261138 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3a60c30-b568-4927-bbe0-fcebead290cb" containerName="mariadb-account-create-update" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.261147 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d367827-344f-49b6-9b67-9e7b08d2c89f" containerName="mariadb-account-create-update" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.261159 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="abb709fd-4b15-4b57-bebd-2e1f945c4334" containerName="mariadb-database-create" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.261167 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="897f5cd4-7cba-452f-bf48-ebe6c3333b76" containerName="mariadb-database-create" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.261643 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.269264 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.269588 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.270129 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.270378 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.270482 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bwpgd" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.293943 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-r5sgw"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.303665 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-28zxv"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.312478 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.327124 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-28zxv"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.375072 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxj4v\" (UniqueName: \"kubernetes.io/projected/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-kube-api-access-qxj4v\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.375129 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-credential-keys\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.375151 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-fernet-keys\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.375194 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-scripts\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.375236 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-combined-ca-bundle\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.375294 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-config-data\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.412072 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-655c5f955f-jh7hn"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.413336 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.415407 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.415475 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-2dfrq" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.419636 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.419940 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.441395 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-655c5f955f-jh7hn"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.478407 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxj4v\" (UniqueName: \"kubernetes.io/projected/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-kube-api-access-qxj4v\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.478447 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.478472 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjsbf\" (UniqueName: \"kubernetes.io/projected/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-kube-api-access-vjsbf\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.478538 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-credential-keys\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.478558 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-fernet-keys\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.478585 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-config\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.478612 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-scripts\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.478651 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-combined-ca-bundle\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.478674 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.478708 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-config-data\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.478733 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.488538 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-credential-keys\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.493034 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-fernet-keys\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.493382 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-combined-ca-bundle\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.504188 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-config-data\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.505364 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxj4v\" (UniqueName: \"kubernetes.io/projected/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-kube-api-access-qxj4v\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.507117 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-scripts\") pod \"keystone-bootstrap-r5sgw\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.535377 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-dn48v"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.536607 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.541531 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7lgn7" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.541745 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.541908 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.553040 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-26gvb"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.554959 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-26gvb" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.556721 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.558838 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jntj9" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.559025 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.582369 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.582469 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39f46125-3c80-4fa1-98d2-88155d268fab-scripts\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.582520 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/39f46125-3c80-4fa1-98d2-88155d268fab-horizon-secret-key\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.582599 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.582862 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.582916 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjsbf\" (UniqueName: \"kubernetes.io/projected/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-kube-api-access-vjsbf\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.582986 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-config\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.589761 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39f46125-3c80-4fa1-98d2-88155d268fab-logs\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.589875 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxbh9\" (UniqueName: \"kubernetes.io/projected/39f46125-3c80-4fa1-98d2-88155d268fab-kube-api-access-xxbh9\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.589942 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39f46125-3c80-4fa1-98d2-88155d268fab-config-data\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.591068 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.591756 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.592341 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.592777 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-config\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.597150 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-dn48v"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.614488 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.631840 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-26gvb"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.647875 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjsbf\" (UniqueName: \"kubernetes.io/projected/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-kube-api-access-vjsbf\") pod \"dnsmasq-dns-5c9d85d47c-28zxv\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.666036 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-28zxv"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.667310 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.704859 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-z2cs6"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.729939 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.732186 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a72a028-d521-466b-9924-eba4595e3750-combined-ca-bundle\") pod \"neutron-db-sync-26gvb\" (UID: \"2a72a028-d521-466b-9924-eba4595e3750\") " pod="openstack/neutron-db-sync-26gvb" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.732261 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39f46125-3c80-4fa1-98d2-88155d268fab-logs\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.732302 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a72a028-d521-466b-9924-eba4595e3750-config\") pod \"neutron-db-sync-26gvb\" (UID: \"2a72a028-d521-466b-9924-eba4595e3750\") " pod="openstack/neutron-db-sync-26gvb" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.732330 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-combined-ca-bundle\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.732355 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-scripts\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.732389 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02ce26c3-6755-479e-93c4-295ef8a3ba2a-etc-machine-id\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.732411 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxbh9\" (UniqueName: \"kubernetes.io/projected/39f46125-3c80-4fa1-98d2-88155d268fab-kube-api-access-xxbh9\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.732467 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39f46125-3c80-4fa1-98d2-88155d268fab-config-data\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.732618 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39f46125-3c80-4fa1-98d2-88155d268fab-scripts\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.732663 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/39f46125-3c80-4fa1-98d2-88155d268fab-horizon-secret-key\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.732727 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-config-data\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.732747 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4667\" (UniqueName: \"kubernetes.io/projected/02ce26c3-6755-479e-93c4-295ef8a3ba2a-kube-api-access-m4667\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.733025 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-db-sync-config-data\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.733123 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29pst\" (UniqueName: \"kubernetes.io/projected/2a72a028-d521-466b-9924-eba4595e3750-kube-api-access-29pst\") pod \"neutron-db-sync-26gvb\" (UID: \"2a72a028-d521-466b-9924-eba4595e3750\") " pod="openstack/neutron-db-sync-26gvb" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.742405 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39f46125-3c80-4fa1-98d2-88155d268fab-logs\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.749181 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39f46125-3c80-4fa1-98d2-88155d268fab-config-data\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.759688 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-vjrqg"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.761086 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.761317 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.761585 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xdxfw" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.762321 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.764928 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/39f46125-3c80-4fa1-98d2-88155d268fab-horizon-secret-key\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.765263 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39f46125-3c80-4fa1-98d2-88155d268fab-scripts\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.765783 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxbh9\" (UniqueName: \"kubernetes.io/projected/39f46125-3c80-4fa1-98d2-88155d268fab-kube-api-access-xxbh9\") pod \"horizon-655c5f955f-jh7hn\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.775987 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-z2cs6"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.811797 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-vjrqg"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.834773 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29pst\" (UniqueName: \"kubernetes.io/projected/2a72a028-d521-466b-9924-eba4595e3750-kube-api-access-29pst\") pod \"neutron-db-sync-26gvb\" (UID: \"2a72a028-d521-466b-9924-eba4595e3750\") " pod="openstack/neutron-db-sync-26gvb" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.834829 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a72a028-d521-466b-9924-eba4595e3750-combined-ca-bundle\") pod \"neutron-db-sync-26gvb\" (UID: \"2a72a028-d521-466b-9924-eba4595e3750\") " pod="openstack/neutron-db-sync-26gvb" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.834849 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-combined-ca-bundle\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.834872 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-config\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.834893 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a72a028-d521-466b-9924-eba4595e3750-config\") pod \"neutron-db-sync-26gvb\" (UID: \"2a72a028-d521-466b-9924-eba4595e3750\") " pod="openstack/neutron-db-sync-26gvb" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.834911 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.834931 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-combined-ca-bundle\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.834949 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-config-data\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.834963 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-scripts\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.834978 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02ce26c3-6755-479e-93c4-295ef8a3ba2a-etc-machine-id\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.834995 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l42th\" (UniqueName: \"kubernetes.io/projected/6041acf3-4620-48a5-8455-6c8cdb50cb0b-kube-api-access-l42th\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.835016 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6041acf3-4620-48a5-8455-6c8cdb50cb0b-logs\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.835032 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-scripts\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.835068 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-dns-svc\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.835093 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.835129 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grt4p\" (UniqueName: \"kubernetes.io/projected/8029d5ff-62c3-4088-8122-1e87a2773995-kube-api-access-grt4p\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.835161 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-config-data\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.835178 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4667\" (UniqueName: \"kubernetes.io/projected/02ce26c3-6755-479e-93c4-295ef8a3ba2a-kube-api-access-m4667\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.835201 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-db-sync-config-data\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.839789 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-scripts\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.842533 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a72a028-d521-466b-9924-eba4595e3750-combined-ca-bundle\") pod \"neutron-db-sync-26gvb\" (UID: \"2a72a028-d521-466b-9924-eba4595e3750\") " pod="openstack/neutron-db-sync-26gvb" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.844766 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02ce26c3-6755-479e-93c4-295ef8a3ba2a-etc-machine-id\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.845952 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a72a028-d521-466b-9924-eba4595e3750-config\") pod \"neutron-db-sync-26gvb\" (UID: \"2a72a028-d521-466b-9924-eba4595e3750\") " pod="openstack/neutron-db-sync-26gvb" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.850047 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-combined-ca-bundle\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.852687 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-db-sync-config-data\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.860351 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.860833 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-config-data\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.863373 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.866248 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.866499 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.868859 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4667\" (UniqueName: \"kubernetes.io/projected/02ce26c3-6755-479e-93c4-295ef8a3ba2a-kube-api-access-m4667\") pod \"cinder-db-sync-dn48v\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.883726 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-pml9v"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.884722 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pml9v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.889677 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29pst\" (UniqueName: \"kubernetes.io/projected/2a72a028-d521-466b-9924-eba4595e3750-kube-api-access-29pst\") pod \"neutron-db-sync-26gvb\" (UID: \"2a72a028-d521-466b-9924-eba4595e3750\") " pod="openstack/neutron-db-sync-26gvb" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.889846 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-7wlxl" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.890055 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.925390 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pml9v"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.933902 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.934422 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dn48v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.935856 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-config-data\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.935884 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l42th\" (UniqueName: \"kubernetes.io/projected/6041acf3-4620-48a5-8455-6c8cdb50cb0b-kube-api-access-l42th\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.935903 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsl75\" (UniqueName: \"kubernetes.io/projected/27762870-383b-4e1b-8d28-e6e7c96d635a-kube-api-access-dsl75\") pod \"barbican-db-sync-pml9v\" (UID: \"27762870-383b-4e1b-8d28-e6e7c96d635a\") " pod="openstack/barbican-db-sync-pml9v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.935919 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctvl6\" (UniqueName: \"kubernetes.io/projected/57b51201-4e02-4b29-843d-6bfe84247cd8-kube-api-access-ctvl6\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.935940 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6041acf3-4620-48a5-8455-6c8cdb50cb0b-logs\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.935957 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-scripts\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.935973 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-scripts\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.935997 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/27762870-383b-4e1b-8d28-e6e7c96d635a-db-sync-config-data\") pod \"barbican-db-sync-pml9v\" (UID: \"27762870-383b-4e1b-8d28-e6e7c96d635a\") " pod="openstack/barbican-db-sync-pml9v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.936017 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-dns-svc\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.936032 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27762870-383b-4e1b-8d28-e6e7c96d635a-combined-ca-bundle\") pod \"barbican-db-sync-pml9v\" (UID: \"27762870-383b-4e1b-8d28-e6e7c96d635a\") " pod="openstack/barbican-db-sync-pml9v" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.936051 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.936082 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grt4p\" (UniqueName: \"kubernetes.io/projected/8029d5ff-62c3-4088-8122-1e87a2773995-kube-api-access-grt4p\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.936106 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b51201-4e02-4b29-843d-6bfe84247cd8-run-httpd\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.936128 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.936145 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b51201-4e02-4b29-843d-6bfe84247cd8-log-httpd\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.936164 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-config-data\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.936184 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.936231 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-combined-ca-bundle\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.936249 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-config\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.936271 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.937210 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.940093 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6041acf3-4620-48a5-8455-6c8cdb50cb0b-logs\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.942942 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.943442 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-config-data\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.944004 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-config\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.944050 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-b697487c7-wzjcs"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.944508 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-dns-svc\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.946170 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-scripts\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.947037 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-combined-ca-bundle\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.949140 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.952515 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-b697487c7-wzjcs"] Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.957869 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l42th\" (UniqueName: \"kubernetes.io/projected/6041acf3-4620-48a5-8455-6c8cdb50cb0b-kube-api-access-l42th\") pod \"placement-db-sync-z2cs6\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:32 crc kubenswrapper[4931]: I1129 00:18:32.967435 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grt4p\" (UniqueName: \"kubernetes.io/projected/8029d5ff-62c3-4088-8122-1e87a2773995-kube-api-access-grt4p\") pod \"dnsmasq-dns-6ffb94d8ff-vjrqg\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038174 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-config-data\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038212 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsl75\" (UniqueName: \"kubernetes.io/projected/27762870-383b-4e1b-8d28-e6e7c96d635a-kube-api-access-dsl75\") pod \"barbican-db-sync-pml9v\" (UID: \"27762870-383b-4e1b-8d28-e6e7c96d635a\") " pod="openstack/barbican-db-sync-pml9v" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038231 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctvl6\" (UniqueName: \"kubernetes.io/projected/57b51201-4e02-4b29-843d-6bfe84247cd8-kube-api-access-ctvl6\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038248 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llfvm\" (UniqueName: \"kubernetes.io/projected/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-kube-api-access-llfvm\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038276 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-scripts\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038294 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-logs\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038316 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/27762870-383b-4e1b-8d28-e6e7c96d635a-db-sync-config-data\") pod \"barbican-db-sync-pml9v\" (UID: \"27762870-383b-4e1b-8d28-e6e7c96d635a\") " pod="openstack/barbican-db-sync-pml9v" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038333 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27762870-383b-4e1b-8d28-e6e7c96d635a-combined-ca-bundle\") pod \"barbican-db-sync-pml9v\" (UID: \"27762870-383b-4e1b-8d28-e6e7c96d635a\") " pod="openstack/barbican-db-sync-pml9v" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038383 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b51201-4e02-4b29-843d-6bfe84247cd8-run-httpd\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038398 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038415 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b51201-4e02-4b29-843d-6bfe84247cd8-log-httpd\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038432 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-config-data\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038450 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038480 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-scripts\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.038505 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-horizon-secret-key\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.043290 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b51201-4e02-4b29-843d-6bfe84247cd8-run-httpd\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.049763 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b51201-4e02-4b29-843d-6bfe84247cd8-log-httpd\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.049777 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.052285 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27762870-383b-4e1b-8d28-e6e7c96d635a-combined-ca-bundle\") pod \"barbican-db-sync-pml9v\" (UID: \"27762870-383b-4e1b-8d28-e6e7c96d635a\") " pod="openstack/barbican-db-sync-pml9v" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.052852 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-scripts\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.052948 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.053522 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/27762870-383b-4e1b-8d28-e6e7c96d635a-db-sync-config-data\") pod \"barbican-db-sync-pml9v\" (UID: \"27762870-383b-4e1b-8d28-e6e7c96d635a\") " pod="openstack/barbican-db-sync-pml9v" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.053656 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.054047 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-config-data\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.057860 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctvl6\" (UniqueName: \"kubernetes.io/projected/57b51201-4e02-4b29-843d-6bfe84247cd8-kube-api-access-ctvl6\") pod \"ceilometer-0\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.060304 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsl75\" (UniqueName: \"kubernetes.io/projected/27762870-383b-4e1b-8d28-e6e7c96d635a-kube-api-access-dsl75\") pod \"barbican-db-sync-pml9v\" (UID: \"27762870-383b-4e1b-8d28-e6e7c96d635a\") " pod="openstack/barbican-db-sync-pml9v" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.087390 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-26gvb" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.094271 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z2cs6" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.117244 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.142234 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-scripts\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.142322 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-horizon-secret-key\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.142385 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-config-data\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.142412 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llfvm\" (UniqueName: \"kubernetes.io/projected/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-kube-api-access-llfvm\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.142459 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-logs\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.142924 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-logs\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.144276 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-config-data\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.144825 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-scripts\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.146307 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-horizon-secret-key\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.163916 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llfvm\" (UniqueName: \"kubernetes.io/projected/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-kube-api-access-llfvm\") pod \"horizon-b697487c7-wzjcs\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.200999 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.280245 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-r5sgw"] Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.284883 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pml9v" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.313401 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.387874 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-28zxv"] Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.468011 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-dn48v"] Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.738317 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-655c5f955f-jh7hn"] Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.972146 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-z2cs6"] Nov 29 00:18:33 crc kubenswrapper[4931]: I1129 00:18:33.982411 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-26gvb"] Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.003359 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-vjrqg"] Nov 29 00:18:34 crc kubenswrapper[4931]: W1129 00:18:34.013827 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8029d5ff_62c3_4088_8122_1e87a2773995.slice/crio-9d80cb1588ece603329894756695d8faa7931d6bf8e20fb381a0e17edf399e94 WatchSource:0}: Error finding container 9d80cb1588ece603329894756695d8faa7931d6bf8e20fb381a0e17edf399e94: Status 404 returned error can't find the container with id 9d80cb1588ece603329894756695d8faa7931d6bf8e20fb381a0e17edf399e94 Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.035943 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-26gvb" event={"ID":"2a72a028-d521-466b-9924-eba4595e3750","Type":"ContainerStarted","Data":"ff120713d04c5d220e419616111b8ea976e8640c850d963a65faa18503643bc5"} Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.072128 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" event={"ID":"8029d5ff-62c3-4088-8122-1e87a2773995","Type":"ContainerStarted","Data":"9d80cb1588ece603329894756695d8faa7931d6bf8e20fb381a0e17edf399e94"} Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.074096 4931 generic.go:334] "Generic (PLEG): container finished" podID="ee6e9ed0-1ced-4049-8a39-ed312c99fce1" containerID="330684c1b0553984f25475ee5db6d4e3f8b97f2c02d2a9fc89c92aa3b800bf6d" exitCode=0 Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.074241 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" event={"ID":"ee6e9ed0-1ced-4049-8a39-ed312c99fce1","Type":"ContainerDied","Data":"330684c1b0553984f25475ee5db6d4e3f8b97f2c02d2a9fc89c92aa3b800bf6d"} Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.074259 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" event={"ID":"ee6e9ed0-1ced-4049-8a39-ed312c99fce1","Type":"ContainerStarted","Data":"ebf16f698f7bc918545aeaf27d6f0566d9cf2620b47efbbdbe05698548f96b5a"} Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.078081 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r5sgw" event={"ID":"5888ab52-af9b-4cf3-a1b6-285f13ffa05e","Type":"ContainerStarted","Data":"fca4eb12fdb64a7cbd05f5848df54f0703027b9db8e4943267818e18d15a5ca0"} Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.078137 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r5sgw" event={"ID":"5888ab52-af9b-4cf3-a1b6-285f13ffa05e","Type":"ContainerStarted","Data":"b47d78d856b456f62bb21f188df4c3722593aa0e272b6e43cf5e6fcd374bd23c"} Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.080734 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z2cs6" event={"ID":"6041acf3-4620-48a5-8455-6c8cdb50cb0b","Type":"ContainerStarted","Data":"9c1a6439c2697030aad62ca0f75ba2cd5d9e6bb1987b4c6a75e193c34da587d3"} Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.084126 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655c5f955f-jh7hn" event={"ID":"39f46125-3c80-4fa1-98d2-88155d268fab","Type":"ContainerStarted","Data":"3ca636d89c167be0e9b5a09f3a2781c7ab78c3b260c42cdaf42ef57ee474311c"} Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.088881 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dn48v" event={"ID":"02ce26c3-6755-479e-93c4-295ef8a3ba2a","Type":"ContainerStarted","Data":"fafbb22a079ec990b25ece1736036e5e20b70806cd25a05ad9895d314a16af09"} Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.108981 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-r5sgw" podStartSLOduration=2.108958352 podStartE2EDuration="2.108958352s" podCreationTimestamp="2025-11-29 00:18:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:18:34.107109169 +0000 UTC m=+1127.269002421" watchObservedRunningTime="2025-11-29 00:18:34.108958352 +0000 UTC m=+1127.270851594" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.189341 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.237311 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pml9v"] Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.266104 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-b697487c7-wzjcs"] Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.445304 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.490176 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-config\") pod \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.490262 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-dns-svc\") pod \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.490479 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjsbf\" (UniqueName: \"kubernetes.io/projected/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-kube-api-access-vjsbf\") pod \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.490521 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-ovsdbserver-sb\") pod \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.490566 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-ovsdbserver-nb\") pod \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\" (UID: \"ee6e9ed0-1ced-4049-8a39-ed312c99fce1\") " Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.504644 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-kube-api-access-vjsbf" (OuterVolumeSpecName: "kube-api-access-vjsbf") pod "ee6e9ed0-1ced-4049-8a39-ed312c99fce1" (UID: "ee6e9ed0-1ced-4049-8a39-ed312c99fce1"). InnerVolumeSpecName "kube-api-access-vjsbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.512391 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ee6e9ed0-1ced-4049-8a39-ed312c99fce1" (UID: "ee6e9ed0-1ced-4049-8a39-ed312c99fce1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.527526 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ee6e9ed0-1ced-4049-8a39-ed312c99fce1" (UID: "ee6e9ed0-1ced-4049-8a39-ed312c99fce1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.542755 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-config" (OuterVolumeSpecName: "config") pod "ee6e9ed0-1ced-4049-8a39-ed312c99fce1" (UID: "ee6e9ed0-1ced-4049-8a39-ed312c99fce1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.561355 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ee6e9ed0-1ced-4049-8a39-ed312c99fce1" (UID: "ee6e9ed0-1ced-4049-8a39-ed312c99fce1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.591739 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.591767 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.591776 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.591786 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjsbf\" (UniqueName: \"kubernetes.io/projected/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-kube-api-access-vjsbf\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.591796 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee6e9ed0-1ced-4049-8a39-ed312c99fce1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.772346 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-655c5f955f-jh7hn"] Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.813598 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-784c4db8c5-j7sst"] Nov 29 00:18:34 crc kubenswrapper[4931]: E1129 00:18:34.814364 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee6e9ed0-1ced-4049-8a39-ed312c99fce1" containerName="init" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.814467 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6e9ed0-1ced-4049-8a39-ed312c99fce1" containerName="init" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.814745 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee6e9ed0-1ced-4049-8a39-ed312c99fce1" containerName="init" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.815976 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.863869 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-784c4db8c5-j7sst"] Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.901582 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.906360 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/653aa9fc-902d-4186-812f-168b8a822c89-horizon-secret-key\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.906476 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwtwm\" (UniqueName: \"kubernetes.io/projected/653aa9fc-902d-4186-812f-168b8a822c89-kube-api-access-dwtwm\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.906501 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/653aa9fc-902d-4186-812f-168b8a822c89-config-data\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.906553 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/653aa9fc-902d-4186-812f-168b8a822c89-scripts\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:34 crc kubenswrapper[4931]: I1129 00:18:34.906582 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/653aa9fc-902d-4186-812f-168b8a822c89-logs\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.008012 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwtwm\" (UniqueName: \"kubernetes.io/projected/653aa9fc-902d-4186-812f-168b8a822c89-kube-api-access-dwtwm\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.008079 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/653aa9fc-902d-4186-812f-168b8a822c89-config-data\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.008130 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/653aa9fc-902d-4186-812f-168b8a822c89-scripts\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.008161 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/653aa9fc-902d-4186-812f-168b8a822c89-logs\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.008237 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/653aa9fc-902d-4186-812f-168b8a822c89-horizon-secret-key\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.010772 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/653aa9fc-902d-4186-812f-168b8a822c89-config-data\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.012181 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/653aa9fc-902d-4186-812f-168b8a822c89-logs\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.012218 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/653aa9fc-902d-4186-812f-168b8a822c89-scripts\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.018569 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/653aa9fc-902d-4186-812f-168b8a822c89-horizon-secret-key\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.030431 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwtwm\" (UniqueName: \"kubernetes.io/projected/653aa9fc-902d-4186-812f-168b8a822c89-kube-api-access-dwtwm\") pod \"horizon-784c4db8c5-j7sst\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.098283 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pml9v" event={"ID":"27762870-383b-4e1b-8d28-e6e7c96d635a","Type":"ContainerStarted","Data":"bb357b079f0955d44569c9ed85ab2a66f066f72acc2320672b75a559cf0ba818"} Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.100605 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-26gvb" event={"ID":"2a72a028-d521-466b-9924-eba4595e3750","Type":"ContainerStarted","Data":"290aa1bf1bb68db7495e27dd7559d5ed0fdf5c812761bb0c24d52ba4dcb49598"} Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.104442 4931 generic.go:334] "Generic (PLEG): container finished" podID="8029d5ff-62c3-4088-8122-1e87a2773995" containerID="a0b67bbd7e690ad64f6bcb71b626d3bde02533eea0236cf86f4f8ed8e1a57775" exitCode=0 Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.104645 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" event={"ID":"8029d5ff-62c3-4088-8122-1e87a2773995","Type":"ContainerDied","Data":"a0b67bbd7e690ad64f6bcb71b626d3bde02533eea0236cf86f4f8ed8e1a57775"} Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.110665 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" event={"ID":"ee6e9ed0-1ced-4049-8a39-ed312c99fce1","Type":"ContainerDied","Data":"ebf16f698f7bc918545aeaf27d6f0566d9cf2620b47efbbdbe05698548f96b5a"} Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.110734 4931 scope.go:117] "RemoveContainer" containerID="330684c1b0553984f25475ee5db6d4e3f8b97f2c02d2a9fc89c92aa3b800bf6d" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.110881 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-28zxv" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.125986 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-26gvb" podStartSLOduration=3.125964863 podStartE2EDuration="3.125964863s" podCreationTimestamp="2025-11-29 00:18:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:18:35.125697966 +0000 UTC m=+1128.287591198" watchObservedRunningTime="2025-11-29 00:18:35.125964863 +0000 UTC m=+1128.287858095" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.155211 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57b51201-4e02-4b29-843d-6bfe84247cd8","Type":"ContainerStarted","Data":"8252cadc8a68ffd1909427114d2e0c176e5f5bd1ac9b013a606a44d9d74ee88f"} Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.157529 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.167561 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b697487c7-wzjcs" event={"ID":"8826b2d2-afb8-449c-8b42-c3d8e50cfe59","Type":"ContainerStarted","Data":"9d18cfef807639acf2ba2c074599af59d96a4b5b216ada3ad40ff7aa4e4d01c6"} Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.270643 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-28zxv"] Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.270712 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-28zxv"] Nov 29 00:18:35 crc kubenswrapper[4931]: I1129 00:18:35.736447 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-784c4db8c5-j7sst"] Nov 29 00:18:35 crc kubenswrapper[4931]: W1129 00:18:35.740031 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod653aa9fc_902d_4186_812f_168b8a822c89.slice/crio-8a412fa3d7d9f02049897ff8a78284ffa73490847c10c37d5f52daa98443db75 WatchSource:0}: Error finding container 8a412fa3d7d9f02049897ff8a78284ffa73490847c10c37d5f52daa98443db75: Status 404 returned error can't find the container with id 8a412fa3d7d9f02049897ff8a78284ffa73490847c10c37d5f52daa98443db75 Nov 29 00:18:36 crc kubenswrapper[4931]: I1129 00:18:36.174132 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-784c4db8c5-j7sst" event={"ID":"653aa9fc-902d-4186-812f-168b8a822c89","Type":"ContainerStarted","Data":"8a412fa3d7d9f02049897ff8a78284ffa73490847c10c37d5f52daa98443db75"} Nov 29 00:18:37 crc kubenswrapper[4931]: I1129 00:18:37.222765 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee6e9ed0-1ced-4049-8a39-ed312c99fce1" path="/var/lib/kubelet/pods/ee6e9ed0-1ced-4049-8a39-ed312c99fce1/volumes" Nov 29 00:18:39 crc kubenswrapper[4931]: I1129 00:18:39.207751 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" event={"ID":"8029d5ff-62c3-4088-8122-1e87a2773995","Type":"ContainerStarted","Data":"c52a3a4e3431c27c0e822288f03d88e131f6e68773372dd961028493a848df85"} Nov 29 00:18:39 crc kubenswrapper[4931]: I1129 00:18:39.208255 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:39 crc kubenswrapper[4931]: I1129 00:18:39.231401 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" podStartSLOduration=7.231382929 podStartE2EDuration="7.231382929s" podCreationTimestamp="2025-11-29 00:18:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:18:39.224593165 +0000 UTC m=+1132.386486407" watchObservedRunningTime="2025-11-29 00:18:39.231382929 +0000 UTC m=+1132.393276161" Nov 29 00:18:40 crc kubenswrapper[4931]: I1129 00:18:40.222506 4931 generic.go:334] "Generic (PLEG): container finished" podID="5888ab52-af9b-4cf3-a1b6-285f13ffa05e" containerID="fca4eb12fdb64a7cbd05f5848df54f0703027b9db8e4943267818e18d15a5ca0" exitCode=0 Nov 29 00:18:40 crc kubenswrapper[4931]: I1129 00:18:40.222660 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r5sgw" event={"ID":"5888ab52-af9b-4cf3-a1b6-285f13ffa05e","Type":"ContainerDied","Data":"fca4eb12fdb64a7cbd05f5848df54f0703027b9db8e4943267818e18d15a5ca0"} Nov 29 00:18:40 crc kubenswrapper[4931]: I1129 00:18:40.985460 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-b697487c7-wzjcs"] Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.013059 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-765c4d856b-6tkcn"] Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.018357 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.023466 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.032291 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-765c4d856b-6tkcn"] Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.034117 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sljht\" (UniqueName: \"kubernetes.io/projected/77f47a1f-3302-4d58-9b94-4ea756cc38ab-kube-api-access-sljht\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.034400 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-horizon-secret-key\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.036585 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-combined-ca-bundle\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.036805 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f47a1f-3302-4d58-9b94-4ea756cc38ab-logs\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.036873 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77f47a1f-3302-4d58-9b94-4ea756cc38ab-scripts\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.036911 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77f47a1f-3302-4d58-9b94-4ea756cc38ab-config-data\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.036958 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-horizon-tls-certs\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.103754 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-784c4db8c5-j7sst"] Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.131752 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79d99cf788-wwqzd"] Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.133340 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.142155 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79d99cf788-wwqzd"] Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.145947 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-combined-ca-bundle\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.146102 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f47a1f-3302-4d58-9b94-4ea756cc38ab-logs\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.146140 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77f47a1f-3302-4d58-9b94-4ea756cc38ab-scripts\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.146169 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77f47a1f-3302-4d58-9b94-4ea756cc38ab-config-data\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.146208 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-horizon-tls-certs\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.146373 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sljht\" (UniqueName: \"kubernetes.io/projected/77f47a1f-3302-4d58-9b94-4ea756cc38ab-kube-api-access-sljht\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.146406 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-horizon-secret-key\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.146571 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f47a1f-3302-4d58-9b94-4ea756cc38ab-logs\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.147959 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77f47a1f-3302-4d58-9b94-4ea756cc38ab-scripts\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.147970 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77f47a1f-3302-4d58-9b94-4ea756cc38ab-config-data\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.163733 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-horizon-secret-key\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.169442 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-combined-ca-bundle\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.190402 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-horizon-tls-certs\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.193426 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sljht\" (UniqueName: \"kubernetes.io/projected/77f47a1f-3302-4d58-9b94-4ea756cc38ab-kube-api-access-sljht\") pod \"horizon-765c4d856b-6tkcn\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.250636 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4003b615-a779-45a5-b95b-913c80226f4a-logs\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.252092 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4003b615-a779-45a5-b95b-913c80226f4a-combined-ca-bundle\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.252195 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4003b615-a779-45a5-b95b-913c80226f4a-config-data\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.252228 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4003b615-a779-45a5-b95b-913c80226f4a-horizon-tls-certs\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.252268 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4003b615-a779-45a5-b95b-913c80226f4a-scripts\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.252333 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4gzn\" (UniqueName: \"kubernetes.io/projected/4003b615-a779-45a5-b95b-913c80226f4a-kube-api-access-j4gzn\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.252383 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4003b615-a779-45a5-b95b-913c80226f4a-horizon-secret-key\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.343234 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.354245 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4003b615-a779-45a5-b95b-913c80226f4a-config-data\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.354285 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4003b615-a779-45a5-b95b-913c80226f4a-horizon-tls-certs\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.354320 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4003b615-a779-45a5-b95b-913c80226f4a-scripts\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.354345 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4gzn\" (UniqueName: \"kubernetes.io/projected/4003b615-a779-45a5-b95b-913c80226f4a-kube-api-access-j4gzn\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.354369 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4003b615-a779-45a5-b95b-913c80226f4a-horizon-secret-key\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.354397 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4003b615-a779-45a5-b95b-913c80226f4a-logs\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.354510 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4003b615-a779-45a5-b95b-913c80226f4a-combined-ca-bundle\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.355409 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4003b615-a779-45a5-b95b-913c80226f4a-config-data\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.355665 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4003b615-a779-45a5-b95b-913c80226f4a-logs\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.356161 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4003b615-a779-45a5-b95b-913c80226f4a-scripts\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.358123 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4003b615-a779-45a5-b95b-913c80226f4a-combined-ca-bundle\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.359444 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4003b615-a779-45a5-b95b-913c80226f4a-horizon-secret-key\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.369318 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4gzn\" (UniqueName: \"kubernetes.io/projected/4003b615-a779-45a5-b95b-913c80226f4a-kube-api-access-j4gzn\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.374829 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4003b615-a779-45a5-b95b-913c80226f4a-horizon-tls-certs\") pod \"horizon-79d99cf788-wwqzd\" (UID: \"4003b615-a779-45a5-b95b-913c80226f4a\") " pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:41 crc kubenswrapper[4931]: I1129 00:18:41.455461 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:18:42 crc kubenswrapper[4931]: I1129 00:18:42.246747 4931 generic.go:334] "Generic (PLEG): container finished" podID="2c957f2c-45cc-4a11-8019-e20de202266b" containerID="549aecf595ecff89023b1a2b77d9b714fa543e1a33c0e1a53ad670b5080a7092" exitCode=0 Nov 29 00:18:42 crc kubenswrapper[4931]: I1129 00:18:42.246799 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-dj2pb" event={"ID":"2c957f2c-45cc-4a11-8019-e20de202266b","Type":"ContainerDied","Data":"549aecf595ecff89023b1a2b77d9b714fa543e1a33c0e1a53ad670b5080a7092"} Nov 29 00:18:43 crc kubenswrapper[4931]: I1129 00:18:43.119448 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:18:43 crc kubenswrapper[4931]: I1129 00:18:43.168070 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-z4dhc"] Nov 29 00:18:43 crc kubenswrapper[4931]: I1129 00:18:43.169416 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" podUID="1f406c3b-dbb1-421e-9cee-334414d6089f" containerName="dnsmasq-dns" containerID="cri-o://35db0a6c7af187c4ddc94c6834c37556eaf379e1acccb939989c780e040403c1" gracePeriod=10 Nov 29 00:18:44 crc kubenswrapper[4931]: I1129 00:18:44.270202 4931 generic.go:334] "Generic (PLEG): container finished" podID="1f406c3b-dbb1-421e-9cee-334414d6089f" containerID="35db0a6c7af187c4ddc94c6834c37556eaf379e1acccb939989c780e040403c1" exitCode=0 Nov 29 00:18:44 crc kubenswrapper[4931]: I1129 00:18:44.270550 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" event={"ID":"1f406c3b-dbb1-421e-9cee-334414d6089f","Type":"ContainerDied","Data":"35db0a6c7af187c4ddc94c6834c37556eaf379e1acccb939989c780e040403c1"} Nov 29 00:18:44 crc kubenswrapper[4931]: I1129 00:18:44.981168 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:44 crc kubenswrapper[4931]: I1129 00:18:44.987045 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-dj2pb" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.040702 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-combined-ca-bundle\") pod \"2c957f2c-45cc-4a11-8019-e20de202266b\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.040741 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-scripts\") pod \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.040797 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-combined-ca-bundle\") pod \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.040880 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-credential-keys\") pod \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.040981 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-825k8\" (UniqueName: \"kubernetes.io/projected/2c957f2c-45cc-4a11-8019-e20de202266b-kube-api-access-825k8\") pod \"2c957f2c-45cc-4a11-8019-e20de202266b\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.041005 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-config-data\") pod \"2c957f2c-45cc-4a11-8019-e20de202266b\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.041034 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-config-data\") pod \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.041067 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxj4v\" (UniqueName: \"kubernetes.io/projected/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-kube-api-access-qxj4v\") pod \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.041088 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-db-sync-config-data\") pod \"2c957f2c-45cc-4a11-8019-e20de202266b\" (UID: \"2c957f2c-45cc-4a11-8019-e20de202266b\") " Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.041166 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-fernet-keys\") pod \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\" (UID: \"5888ab52-af9b-4cf3-a1b6-285f13ffa05e\") " Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.048300 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c957f2c-45cc-4a11-8019-e20de202266b-kube-api-access-825k8" (OuterVolumeSpecName: "kube-api-access-825k8") pod "2c957f2c-45cc-4a11-8019-e20de202266b" (UID: "2c957f2c-45cc-4a11-8019-e20de202266b"). InnerVolumeSpecName "kube-api-access-825k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.066086 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-kube-api-access-qxj4v" (OuterVolumeSpecName: "kube-api-access-qxj4v") pod "5888ab52-af9b-4cf3-a1b6-285f13ffa05e" (UID: "5888ab52-af9b-4cf3-a1b6-285f13ffa05e"). InnerVolumeSpecName "kube-api-access-qxj4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.073341 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5888ab52-af9b-4cf3-a1b6-285f13ffa05e" (UID: "5888ab52-af9b-4cf3-a1b6-285f13ffa05e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.075049 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5888ab52-af9b-4cf3-a1b6-285f13ffa05e" (UID: "5888ab52-af9b-4cf3-a1b6-285f13ffa05e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.111087 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-scripts" (OuterVolumeSpecName: "scripts") pod "5888ab52-af9b-4cf3-a1b6-285f13ffa05e" (UID: "5888ab52-af9b-4cf3-a1b6-285f13ffa05e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.144909 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-825k8\" (UniqueName: \"kubernetes.io/projected/2c957f2c-45cc-4a11-8019-e20de202266b-kube-api-access-825k8\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.144942 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxj4v\" (UniqueName: \"kubernetes.io/projected/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-kube-api-access-qxj4v\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.144951 4931 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.144959 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.144967 4931 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.153023 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2c957f2c-45cc-4a11-8019-e20de202266b" (UID: "2c957f2c-45cc-4a11-8019-e20de202266b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.166896 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5888ab52-af9b-4cf3-a1b6-285f13ffa05e" (UID: "5888ab52-af9b-4cf3-a1b6-285f13ffa05e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.187865 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-config-data" (OuterVolumeSpecName: "config-data") pod "5888ab52-af9b-4cf3-a1b6-285f13ffa05e" (UID: "5888ab52-af9b-4cf3-a1b6-285f13ffa05e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.196288 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c957f2c-45cc-4a11-8019-e20de202266b" (UID: "2c957f2c-45cc-4a11-8019-e20de202266b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.216277 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-config-data" (OuterVolumeSpecName: "config-data") pod "2c957f2c-45cc-4a11-8019-e20de202266b" (UID: "2c957f2c-45cc-4a11-8019-e20de202266b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.247159 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.247181 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.247190 4931 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.247200 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c957f2c-45cc-4a11-8019-e20de202266b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.247209 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5888ab52-af9b-4cf3-a1b6-285f13ffa05e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.303825 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-dj2pb" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.303840 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-dj2pb" event={"ID":"2c957f2c-45cc-4a11-8019-e20de202266b","Type":"ContainerDied","Data":"b01a3e25c1ff1510946ca9a777be650d80a7ef87abda1e49055488f4779af556"} Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.304976 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b01a3e25c1ff1510946ca9a777be650d80a7ef87abda1e49055488f4779af556" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.305909 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r5sgw" event={"ID":"5888ab52-af9b-4cf3-a1b6-285f13ffa05e","Type":"ContainerDied","Data":"b47d78d856b456f62bb21f188df4c3722593aa0e272b6e43cf5e6fcd374bd23c"} Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.305934 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b47d78d856b456f62bb21f188df4c3722593aa0e272b6e43cf5e6fcd374bd23c" Nov 29 00:18:45 crc kubenswrapper[4931]: I1129 00:18:45.305979 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r5sgw" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.092277 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-r5sgw"] Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.101170 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-r5sgw"] Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.196386 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-wknnq"] Nov 29 00:18:46 crc kubenswrapper[4931]: E1129 00:18:46.196821 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5888ab52-af9b-4cf3-a1b6-285f13ffa05e" containerName="keystone-bootstrap" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.196838 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="5888ab52-af9b-4cf3-a1b6-285f13ffa05e" containerName="keystone-bootstrap" Nov 29 00:18:46 crc kubenswrapper[4931]: E1129 00:18:46.196887 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c957f2c-45cc-4a11-8019-e20de202266b" containerName="glance-db-sync" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.196896 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c957f2c-45cc-4a11-8019-e20de202266b" containerName="glance-db-sync" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.197095 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="5888ab52-af9b-4cf3-a1b6-285f13ffa05e" containerName="keystone-bootstrap" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.197114 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c957f2c-45cc-4a11-8019-e20de202266b" containerName="glance-db-sync" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.197770 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.201947 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.202236 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.202402 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.202732 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.203251 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bwpgd" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.218124 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wknnq"] Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.263975 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-scripts\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.264079 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-credential-keys\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.264117 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-fernet-keys\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.264172 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-config-data\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.264218 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-combined-ca-bundle\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.264285 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqvf9\" (UniqueName: \"kubernetes.io/projected/8e6ded52-583f-44e7-853c-30c354886ae9-kube-api-access-cqvf9\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.366015 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-scripts\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.366072 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-credential-keys\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.366101 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-fernet-keys\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.366151 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-config-data\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.366174 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-combined-ca-bundle\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.366222 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqvf9\" (UniqueName: \"kubernetes.io/projected/8e6ded52-583f-44e7-853c-30c354886ae9-kube-api-access-cqvf9\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.373238 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-scripts\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.376994 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-credential-keys\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.378872 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-combined-ca-bundle\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.381518 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-fernet-keys\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.382158 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-config-data\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.395447 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqvf9\" (UniqueName: \"kubernetes.io/projected/8e6ded52-583f-44e7-853c-30c354886ae9-kube-api-access-cqvf9\") pod \"keystone-bootstrap-wknnq\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.438564 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56798b757f-jxjxc"] Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.440368 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.467222 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-ovsdbserver-sb\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.467261 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt2kz\" (UniqueName: \"kubernetes.io/projected/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-kube-api-access-dt2kz\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.467304 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-dns-svc\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.467338 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-ovsdbserver-nb\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.467433 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-config\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.468883 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56798b757f-jxjxc"] Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.533538 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.568906 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-config\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.568954 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-ovsdbserver-sb\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.568983 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt2kz\" (UniqueName: \"kubernetes.io/projected/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-kube-api-access-dt2kz\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.569023 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-dns-svc\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.569058 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-ovsdbserver-nb\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.570043 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-ovsdbserver-sb\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.570070 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-dns-svc\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.570104 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-ovsdbserver-nb\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.570112 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-config\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.593293 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt2kz\" (UniqueName: \"kubernetes.io/projected/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-kube-api-access-dt2kz\") pod \"dnsmasq-dns-56798b757f-jxjxc\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:46 crc kubenswrapper[4931]: I1129 00:18:46.788931 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.247621 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5888ab52-af9b-4cf3-a1b6-285f13ffa05e" path="/var/lib/kubelet/pods/5888ab52-af9b-4cf3-a1b6-285f13ffa05e/volumes" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.360635 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.362678 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.366290 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-m5lx4" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.366475 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.366545 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.376028 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.490032 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.490355 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af2b8a70-d017-445e-a7bd-f4b19309abf7-logs\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.490571 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.490689 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-config-data\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.490792 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af2b8a70-d017-445e-a7bd-f4b19309abf7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.490959 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dln4c\" (UniqueName: \"kubernetes.io/projected/af2b8a70-d017-445e-a7bd-f4b19309abf7-kube-api-access-dln4c\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.491057 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-scripts\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.563942 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.565310 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.569013 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.583328 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592054 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592114 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592137 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-config-data\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592160 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af2b8a70-d017-445e-a7bd-f4b19309abf7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592216 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592265 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592286 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-logs\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592332 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dln4c\" (UniqueName: \"kubernetes.io/projected/af2b8a70-d017-445e-a7bd-f4b19309abf7-kube-api-access-dln4c\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592348 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptbb2\" (UniqueName: \"kubernetes.io/projected/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-kube-api-access-ptbb2\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592413 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-scripts\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592532 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592614 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592738 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592785 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af2b8a70-d017-445e-a7bd-f4b19309abf7-logs\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.592929 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.593501 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af2b8a70-d017-445e-a7bd-f4b19309abf7-logs\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.593518 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af2b8a70-d017-445e-a7bd-f4b19309abf7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.599836 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-config-data\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.600575 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.612823 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-scripts\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.616016 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dln4c\" (UniqueName: \"kubernetes.io/projected/af2b8a70-d017-445e-a7bd-f4b19309abf7-kube-api-access-dln4c\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.628485 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.687597 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.694662 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.694743 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.694772 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-logs\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.694830 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptbb2\" (UniqueName: \"kubernetes.io/projected/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-kube-api-access-ptbb2\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.694907 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.694970 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.695028 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.695216 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.695432 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-logs\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.696392 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.700208 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.700570 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.702754 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.714721 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptbb2\" (UniqueName: \"kubernetes.io/projected/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-kube-api-access-ptbb2\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.746025 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.890019 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 00:18:47 crc kubenswrapper[4931]: I1129 00:18:47.898434 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:18:47 crc kubenswrapper[4931]: E1129 00:18:47.898639 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:18:47 crc kubenswrapper[4931]: E1129 00:18:47.898674 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 29 00:18:47 crc kubenswrapper[4931]: E1129 00:18:47.898749 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift podName:a720d879-8b09-452d-8371-9bf75f28cbde nodeName:}" failed. No retries permitted until 2025-11-29 00:19:51.898729983 +0000 UTC m=+1205.060623225 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift") pod "swift-storage-0" (UID: "a720d879-8b09-452d-8371-9bf75f28cbde") : configmap "swift-ring-files" not found Nov 29 00:18:48 crc kubenswrapper[4931]: I1129 00:18:48.830326 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:18:48 crc kubenswrapper[4931]: I1129 00:18:48.904565 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:18:53 crc kubenswrapper[4931]: I1129 00:18:53.083474 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" podUID="1f406c3b-dbb1-421e-9cee-334414d6089f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Nov 29 00:18:57 crc kubenswrapper[4931]: I1129 00:18:57.982163 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.081327 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-ovsdbserver-sb\") pod \"1f406c3b-dbb1-421e-9cee-334414d6089f\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.081778 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-ovsdbserver-nb\") pod \"1f406c3b-dbb1-421e-9cee-334414d6089f\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.081993 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-dns-svc\") pod \"1f406c3b-dbb1-421e-9cee-334414d6089f\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.082039 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-config\") pod \"1f406c3b-dbb1-421e-9cee-334414d6089f\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.082070 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq2k4\" (UniqueName: \"kubernetes.io/projected/1f406c3b-dbb1-421e-9cee-334414d6089f-kube-api-access-zq2k4\") pod \"1f406c3b-dbb1-421e-9cee-334414d6089f\" (UID: \"1f406c3b-dbb1-421e-9cee-334414d6089f\") " Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.084941 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" podUID="1f406c3b-dbb1-421e-9cee-334414d6089f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.091235 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f406c3b-dbb1-421e-9cee-334414d6089f-kube-api-access-zq2k4" (OuterVolumeSpecName: "kube-api-access-zq2k4") pod "1f406c3b-dbb1-421e-9cee-334414d6089f" (UID: "1f406c3b-dbb1-421e-9cee-334414d6089f"). InnerVolumeSpecName "kube-api-access-zq2k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.125700 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-config" (OuterVolumeSpecName: "config") pod "1f406c3b-dbb1-421e-9cee-334414d6089f" (UID: "1f406c3b-dbb1-421e-9cee-334414d6089f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.126421 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1f406c3b-dbb1-421e-9cee-334414d6089f" (UID: "1f406c3b-dbb1-421e-9cee-334414d6089f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.134175 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1f406c3b-dbb1-421e-9cee-334414d6089f" (UID: "1f406c3b-dbb1-421e-9cee-334414d6089f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.151298 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1f406c3b-dbb1-421e-9cee-334414d6089f" (UID: "1f406c3b-dbb1-421e-9cee-334414d6089f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.184659 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.184711 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.184723 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq2k4\" (UniqueName: \"kubernetes.io/projected/1f406c3b-dbb1-421e-9cee-334414d6089f-kube-api-access-zq2k4\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.184739 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.184751 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f406c3b-dbb1-421e-9cee-334414d6089f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.417420 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" event={"ID":"1f406c3b-dbb1-421e-9cee-334414d6089f","Type":"ContainerDied","Data":"4a2764f5bc2907df2f53bd86608548af2afe57f2ab04b6dd9999403af860c296"} Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.417475 4931 scope.go:117] "RemoveContainer" containerID="35db0a6c7af187c4ddc94c6834c37556eaf379e1acccb939989c780e040403c1" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.417602 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-z4dhc" Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.448755 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-z4dhc"] Nov 29 00:18:58 crc kubenswrapper[4931]: I1129 00:18:58.456153 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-z4dhc"] Nov 29 00:18:59 crc kubenswrapper[4931]: I1129 00:18:59.225155 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f406c3b-dbb1-421e-9cee-334414d6089f" path="/var/lib/kubelet/pods/1f406c3b-dbb1-421e-9cee-334414d6089f/volumes" Nov 29 00:19:00 crc kubenswrapper[4931]: E1129 00:19:00.243903 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 29 00:19:00 crc kubenswrapper[4931]: E1129 00:19:00.244307 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m4667,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-dn48v_openstack(02ce26c3-6755-479e-93c4-295ef8a3ba2a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 00:19:00 crc kubenswrapper[4931]: E1129 00:19:00.246310 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-dn48v" podUID="02ce26c3-6755-479e-93c4-295ef8a3ba2a" Nov 29 00:19:00 crc kubenswrapper[4931]: E1129 00:19:00.438237 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-dn48v" podUID="02ce26c3-6755-479e-93c4-295ef8a3ba2a" Nov 29 00:19:00 crc kubenswrapper[4931]: I1129 00:19:00.654046 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79d99cf788-wwqzd"] Nov 29 00:19:01 crc kubenswrapper[4931]: W1129 00:19:01.684600 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4003b615_a779_45a5_b95b_913c80226f4a.slice/crio-acb78bb305595747d3a9a21de89deca1ec76073336d9ad6d57f0f47f72dc471b WatchSource:0}: Error finding container acb78bb305595747d3a9a21de89deca1ec76073336d9ad6d57f0f47f72dc471b: Status 404 returned error can't find the container with id acb78bb305595747d3a9a21de89deca1ec76073336d9ad6d57f0f47f72dc471b Nov 29 00:19:02 crc kubenswrapper[4931]: I1129 00:19:02.088384 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-765c4d856b-6tkcn"] Nov 29 00:19:02 crc kubenswrapper[4931]: I1129 00:19:02.453252 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79d99cf788-wwqzd" event={"ID":"4003b615-a779-45a5-b95b-913c80226f4a","Type":"ContainerStarted","Data":"acb78bb305595747d3a9a21de89deca1ec76073336d9ad6d57f0f47f72dc471b"} Nov 29 00:19:04 crc kubenswrapper[4931]: I1129 00:19:04.155919 4931 scope.go:117] "RemoveContainer" containerID="6013780e650060493eb5a0ffaf010491e43abc96c9a37d3a49a92e3a0a53fdf6" Nov 29 00:19:04 crc kubenswrapper[4931]: W1129 00:19:04.175108 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77f47a1f_3302_4d58_9b94_4ea756cc38ab.slice/crio-87a397bd41a2a0b89231a8081f0bf70144548da1077d3da9e20a53b68b05d55d WatchSource:0}: Error finding container 87a397bd41a2a0b89231a8081f0bf70144548da1077d3da9e20a53b68b05d55d: Status 404 returned error can't find the container with id 87a397bd41a2a0b89231a8081f0bf70144548da1077d3da9e20a53b68b05d55d Nov 29 00:19:04 crc kubenswrapper[4931]: I1129 00:19:04.468304 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-765c4d856b-6tkcn" event={"ID":"77f47a1f-3302-4d58-9b94-4ea756cc38ab","Type":"ContainerStarted","Data":"87a397bd41a2a0b89231a8081f0bf70144548da1077d3da9e20a53b68b05d55d"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.208209 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.242300 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56798b757f-jxjxc"] Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.354431 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wknnq"] Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.415466 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.429307 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.612052 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"af2b8a70-d017-445e-a7bd-f4b19309abf7","Type":"ContainerStarted","Data":"43897677b6508b47fb745cd281b673a69571c5af302ce9c553172b7cbeb86d1b"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.614840 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" event={"ID":"d3c1e52d-f484-4f0c-abbd-2f27a7e83592","Type":"ContainerStarted","Data":"75089c71cb2000d0679a5069803060ed7da1e8910e5654f0e54f26af6da33e9d"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.618060 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655c5f955f-jh7hn" event={"ID":"39f46125-3c80-4fa1-98d2-88155d268fab","Type":"ContainerStarted","Data":"de9118603db7baa81f5d1a801f22d52cc51caf6ed9abfd7300c882443bfe50e0"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.618110 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655c5f955f-jh7hn" event={"ID":"39f46125-3c80-4fa1-98d2-88155d268fab","Type":"ContainerStarted","Data":"1a4a4de6d308cbdb34813d3d4f0f7c6e22ccfeb993fa0be055f679b07a30ac33"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.618150 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-655c5f955f-jh7hn" podUID="39f46125-3c80-4fa1-98d2-88155d268fab" containerName="horizon-log" containerID="cri-o://1a4a4de6d308cbdb34813d3d4f0f7c6e22ccfeb993fa0be055f679b07a30ac33" gracePeriod=30 Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.618178 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-655c5f955f-jh7hn" podUID="39f46125-3c80-4fa1-98d2-88155d268fab" containerName="horizon" containerID="cri-o://de9118603db7baa81f5d1a801f22d52cc51caf6ed9abfd7300c882443bfe50e0" gracePeriod=30 Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.619505 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c","Type":"ContainerStarted","Data":"c383c40c38fccee4291233801ec9208211f95197ad31cde74989ff9feba92b01"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.623167 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57b51201-4e02-4b29-843d-6bfe84247cd8","Type":"ContainerStarted","Data":"bdedfb8b573d44dcd2880bd5f7719721cac9337ba0fe28318bdbf0e5b580ad90"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.631370 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z2cs6" event={"ID":"6041acf3-4620-48a5-8455-6c8cdb50cb0b","Type":"ContainerStarted","Data":"691315ec4d054a167417bbadfd4ddb939e88c1eed51bf22e2853223ea252e573"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.632728 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wknnq" event={"ID":"8e6ded52-583f-44e7-853c-30c354886ae9","Type":"ContainerStarted","Data":"6634eae225d37b06dc6f94e01633c6409d554b410423aaaa9acf665f68a4f6fc"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.642685 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-784c4db8c5-j7sst" event={"ID":"653aa9fc-902d-4186-812f-168b8a822c89","Type":"ContainerStarted","Data":"814dd27c23460a30a37856085da4a4026f6f6a5c5259467c0ea4b7ba5818f758"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.642728 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-784c4db8c5-j7sst" event={"ID":"653aa9fc-902d-4186-812f-168b8a822c89","Type":"ContainerStarted","Data":"d78e686bf2fd86b74ce23a091bdf60c608d88f5d204a9866f5c892ce4b67221b"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.650462 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-655c5f955f-jh7hn" podStartSLOduration=8.255362069 podStartE2EDuration="38.650445078s" podCreationTimestamp="2025-11-29 00:18:32 +0000 UTC" firstStartedPulling="2025-11-29 00:18:33.752077309 +0000 UTC m=+1126.913970541" lastFinishedPulling="2025-11-29 00:19:04.147160318 +0000 UTC m=+1157.309053550" observedRunningTime="2025-11-29 00:19:10.648238775 +0000 UTC m=+1163.810132027" watchObservedRunningTime="2025-11-29 00:19:10.650445078 +0000 UTC m=+1163.812338310" Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.687113 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79d99cf788-wwqzd" event={"ID":"4003b615-a779-45a5-b95b-913c80226f4a","Type":"ContainerStarted","Data":"73efd149a3da817b476d45a88113935b0a1e1483d5f70b1d08c65697aa19cb4e"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.687159 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79d99cf788-wwqzd" event={"ID":"4003b615-a779-45a5-b95b-913c80226f4a","Type":"ContainerStarted","Data":"50d3c604982a53206198e0d053df92fa2ada429a6c28e44bee9b026b43a37216"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.693465 4931 generic.go:334] "Generic (PLEG): container finished" podID="2a72a028-d521-466b-9924-eba4595e3750" containerID="290aa1bf1bb68db7495e27dd7559d5ed0fdf5c812761bb0c24d52ba4dcb49598" exitCode=0 Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.693572 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-26gvb" event={"ID":"2a72a028-d521-466b-9924-eba4595e3750","Type":"ContainerDied","Data":"290aa1bf1bb68db7495e27dd7559d5ed0fdf5c812761bb0c24d52ba4dcb49598"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.702140 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b697487c7-wzjcs" event={"ID":"8826b2d2-afb8-449c-8b42-c3d8e50cfe59","Type":"ContainerStarted","Data":"a58ec12db7bf12bdee74fad4da782c91dc8a2f6eea9b6cf6462f9c481932d58a"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.702177 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b697487c7-wzjcs" event={"ID":"8826b2d2-afb8-449c-8b42-c3d8e50cfe59","Type":"ContainerStarted","Data":"0683ba9e2568040443e68a523b3b1a085df5239d2cd7c5bd50471c51d6091d8a"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.713960 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pml9v" event={"ID":"27762870-383b-4e1b-8d28-e6e7c96d635a","Type":"ContainerStarted","Data":"9a39ab75b839a4c6e4d50a13762239245f3d69a686155574f10f887436ee76e3"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.715398 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-z2cs6" podStartSLOduration=3.000235016 podStartE2EDuration="38.715374116s" podCreationTimestamp="2025-11-29 00:18:32 +0000 UTC" firstStartedPulling="2025-11-29 00:18:34.008698333 +0000 UTC m=+1127.170591575" lastFinishedPulling="2025-11-29 00:19:09.723837443 +0000 UTC m=+1162.885730675" observedRunningTime="2025-11-29 00:19:10.681801896 +0000 UTC m=+1163.843695138" watchObservedRunningTime="2025-11-29 00:19:10.715374116 +0000 UTC m=+1163.877267348" Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.719544 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-765c4d856b-6tkcn" event={"ID":"77f47a1f-3302-4d58-9b94-4ea756cc38ab","Type":"ContainerStarted","Data":"da9b24f68267e8a0f02647614f89605bc826dc9c40f4d64d967492a86d47c0eb"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:10.754735 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-pml9v" podStartSLOduration=3.136789903 podStartE2EDuration="38.754714542s" podCreationTimestamp="2025-11-29 00:18:32 +0000 UTC" firstStartedPulling="2025-11-29 00:18:34.245147249 +0000 UTC m=+1127.407040481" lastFinishedPulling="2025-11-29 00:19:09.863071888 +0000 UTC m=+1163.024965120" observedRunningTime="2025-11-29 00:19:10.734255887 +0000 UTC m=+1163.896149119" watchObservedRunningTime="2025-11-29 00:19:10.754714542 +0000 UTC m=+1163.916607774" Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.760111 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"af2b8a70-d017-445e-a7bd-f4b19309abf7","Type":"ContainerStarted","Data":"99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.764287 4931 generic.go:334] "Generic (PLEG): container finished" podID="d3c1e52d-f484-4f0c-abbd-2f27a7e83592" containerID="91982a165b703336f2c42a80fccbcc2bd30cf39c7246caed00c275a969f81a58" exitCode=0 Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.764346 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" event={"ID":"d3c1e52d-f484-4f0c-abbd-2f27a7e83592","Type":"ContainerDied","Data":"91982a165b703336f2c42a80fccbcc2bd30cf39c7246caed00c275a969f81a58"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.773472 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wknnq" event={"ID":"8e6ded52-583f-44e7-853c-30c354886ae9","Type":"ContainerStarted","Data":"f1ad50d40ba416d98d59313fc595dc27e2284b8f1c20fc2459e688f103bfe16d"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.781939 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c","Type":"ContainerStarted","Data":"351b54a1ca62d1073c7e1968feda227f49aa0edeb9d4376f3ea4419fb1c96403"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.825678 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-b697487c7-wzjcs" podUID="8826b2d2-afb8-449c-8b42-c3d8e50cfe59" containerName="horizon-log" containerID="cri-o://0683ba9e2568040443e68a523b3b1a085df5239d2cd7c5bd50471c51d6091d8a" gracePeriod=30 Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.826162 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-b697487c7-wzjcs" podUID="8826b2d2-afb8-449c-8b42-c3d8e50cfe59" containerName="horizon" containerID="cri-o://a58ec12db7bf12bdee74fad4da782c91dc8a2f6eea9b6cf6462f9c481932d58a" gracePeriod=30 Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.826289 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-765c4d856b-6tkcn" event={"ID":"77f47a1f-3302-4d58-9b94-4ea756cc38ab","Type":"ContainerStarted","Data":"44ea44f6e434cd440fd70136642061dd183226c05674df61c397df90f3874531"} Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.826958 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-784c4db8c5-j7sst" podUID="653aa9fc-902d-4186-812f-168b8a822c89" containerName="horizon-log" containerID="cri-o://d78e686bf2fd86b74ce23a091bdf60c608d88f5d204a9866f5c892ce4b67221b" gracePeriod=30 Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.827432 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-784c4db8c5-j7sst" podUID="653aa9fc-902d-4186-812f-168b8a822c89" containerName="horizon" containerID="cri-o://814dd27c23460a30a37856085da4a4026f6f6a5c5259467c0ea4b7ba5818f758" gracePeriod=30 Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.846425 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-wknnq" podStartSLOduration=25.84639065 podStartE2EDuration="25.84639065s" podCreationTimestamp="2025-11-29 00:18:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:11.826059729 +0000 UTC m=+1164.987952981" watchObservedRunningTime="2025-11-29 00:19:11.84639065 +0000 UTC m=+1165.008283882" Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.951337 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-b697487c7-wzjcs" podStartSLOduration=4.490627553 podStartE2EDuration="39.951312073s" podCreationTimestamp="2025-11-29 00:18:32 +0000 UTC" firstStartedPulling="2025-11-29 00:18:34.256670108 +0000 UTC m=+1127.418563340" lastFinishedPulling="2025-11-29 00:19:09.717354628 +0000 UTC m=+1162.879247860" observedRunningTime="2025-11-29 00:19:11.879715254 +0000 UTC m=+1165.041608496" watchObservedRunningTime="2025-11-29 00:19:11.951312073 +0000 UTC m=+1165.113205325" Nov 29 00:19:11 crc kubenswrapper[4931]: I1129 00:19:11.987634 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-79d99cf788-wwqzd" podStartSLOduration=30.987619062 podStartE2EDuration="30.987619062s" podCreationTimestamp="2025-11-29 00:18:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:11.958409446 +0000 UTC m=+1165.120302698" watchObservedRunningTime="2025-11-29 00:19:11.987619062 +0000 UTC m=+1165.149512294" Nov 29 00:19:12 crc kubenswrapper[4931]: I1129 00:19:12.015262 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-784c4db8c5-j7sst" podStartSLOduration=4.62750683 podStartE2EDuration="38.015234672s" podCreationTimestamp="2025-11-29 00:18:34 +0000 UTC" firstStartedPulling="2025-11-29 00:18:35.743373261 +0000 UTC m=+1128.905266493" lastFinishedPulling="2025-11-29 00:19:09.131101103 +0000 UTC m=+1162.292994335" observedRunningTime="2025-11-29 00:19:11.985026137 +0000 UTC m=+1165.146919369" watchObservedRunningTime="2025-11-29 00:19:12.015234672 +0000 UTC m=+1165.177127904" Nov 29 00:19:12 crc kubenswrapper[4931]: I1129 00:19:12.038235 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-765c4d856b-6tkcn" podStartSLOduration=32.038215879 podStartE2EDuration="32.038215879s" podCreationTimestamp="2025-11-29 00:18:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:12.009709354 +0000 UTC m=+1165.171602596" watchObservedRunningTime="2025-11-29 00:19:12.038215879 +0000 UTC m=+1165.200109111" Nov 29 00:19:12 crc kubenswrapper[4931]: I1129 00:19:12.838937 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" event={"ID":"d3c1e52d-f484-4f0c-abbd-2f27a7e83592","Type":"ContainerStarted","Data":"813b56703bc6e2cfd71c7ba156b1834f10a2c64959025602085c975a22a40b42"} Nov 29 00:19:12 crc kubenswrapper[4931]: I1129 00:19:12.841334 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c","Type":"ContainerStarted","Data":"53fece3204c7e5ba3c8b6b648c62a2dd8b6e51738b70debd3cfdcf677bd17b2f"} Nov 29 00:19:12 crc kubenswrapper[4931]: I1129 00:19:12.844172 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"af2b8a70-d017-445e-a7bd-f4b19309abf7","Type":"ContainerStarted","Data":"7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd"} Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.006029 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-26gvb" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.051336 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.120475 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29pst\" (UniqueName: \"kubernetes.io/projected/2a72a028-d521-466b-9924-eba4595e3750-kube-api-access-29pst\") pod \"2a72a028-d521-466b-9924-eba4595e3750\" (UID: \"2a72a028-d521-466b-9924-eba4595e3750\") " Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.120566 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a72a028-d521-466b-9924-eba4595e3750-combined-ca-bundle\") pod \"2a72a028-d521-466b-9924-eba4595e3750\" (UID: \"2a72a028-d521-466b-9924-eba4595e3750\") " Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.120677 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a72a028-d521-466b-9924-eba4595e3750-config\") pod \"2a72a028-d521-466b-9924-eba4595e3750\" (UID: \"2a72a028-d521-466b-9924-eba4595e3750\") " Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.128855 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a72a028-d521-466b-9924-eba4595e3750-kube-api-access-29pst" (OuterVolumeSpecName: "kube-api-access-29pst") pod "2a72a028-d521-466b-9924-eba4595e3750" (UID: "2a72a028-d521-466b-9924-eba4595e3750"). InnerVolumeSpecName "kube-api-access-29pst". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.152376 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a72a028-d521-466b-9924-eba4595e3750-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a72a028-d521-466b-9924-eba4595e3750" (UID: "2a72a028-d521-466b-9924-eba4595e3750"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.152407 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a72a028-d521-466b-9924-eba4595e3750-config" (OuterVolumeSpecName: "config") pod "2a72a028-d521-466b-9924-eba4595e3750" (UID: "2a72a028-d521-466b-9924-eba4595e3750"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.221902 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29pst\" (UniqueName: \"kubernetes.io/projected/2a72a028-d521-466b-9924-eba4595e3750-kube-api-access-29pst\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.221931 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a72a028-d521-466b-9924-eba4595e3750-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.221939 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a72a028-d521-466b-9924-eba4595e3750-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.314684 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.855074 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-26gvb" event={"ID":"2a72a028-d521-466b-9924-eba4595e3750","Type":"ContainerDied","Data":"ff120713d04c5d220e419616111b8ea976e8640c850d963a65faa18503643bc5"} Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.855370 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff120713d04c5d220e419616111b8ea976e8640c850d963a65faa18503643bc5" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.855422 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-26gvb" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.864900 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57b51201-4e02-4b29-843d-6bfe84247cd8","Type":"ContainerStarted","Data":"8e97f61f77404e3bbfb6af14a2ed78fa4f9311b5b8a206712ad11e9b1b2cb0b4"} Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.865005 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="af2b8a70-d017-445e-a7bd-f4b19309abf7" containerName="glance-log" containerID="cri-o://99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100" gracePeriod=30 Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.865114 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" containerName="glance-log" containerID="cri-o://351b54a1ca62d1073c7e1968feda227f49aa0edeb9d4376f3ea4419fb1c96403" gracePeriod=30 Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.865161 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.865215 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="af2b8a70-d017-445e-a7bd-f4b19309abf7" containerName="glance-httpd" containerID="cri-o://7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd" gracePeriod=30 Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.865379 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" containerName="glance-httpd" containerID="cri-o://53fece3204c7e5ba3c8b6b648c62a2dd8b6e51738b70debd3cfdcf677bd17b2f" gracePeriod=30 Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.887588 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" podStartSLOduration=27.887571565000002 podStartE2EDuration="27.887571565s" podCreationTimestamp="2025-11-29 00:18:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:13.885464745 +0000 UTC m=+1167.047357977" watchObservedRunningTime="2025-11-29 00:19:13.887571565 +0000 UTC m=+1167.049464787" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.916041 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=27.916022943 podStartE2EDuration="27.916022943s" podCreationTimestamp="2025-11-29 00:18:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:13.907487488 +0000 UTC m=+1167.069380730" watchObservedRunningTime="2025-11-29 00:19:13.916022943 +0000 UTC m=+1167.077916175" Nov 29 00:19:13 crc kubenswrapper[4931]: I1129 00:19:13.933150 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=27.933132605 podStartE2EDuration="27.933132605s" podCreationTimestamp="2025-11-29 00:18:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:13.924997081 +0000 UTC m=+1167.086890313" watchObservedRunningTime="2025-11-29 00:19:13.933132605 +0000 UTC m=+1167.095025837" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.289901 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56798b757f-jxjxc"] Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.341690 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b6c948c7-rjvnp"] Nov 29 00:19:14 crc kubenswrapper[4931]: E1129 00:19:14.342174 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f406c3b-dbb1-421e-9cee-334414d6089f" containerName="dnsmasq-dns" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.342198 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f406c3b-dbb1-421e-9cee-334414d6089f" containerName="dnsmasq-dns" Nov 29 00:19:14 crc kubenswrapper[4931]: E1129 00:19:14.342219 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f406c3b-dbb1-421e-9cee-334414d6089f" containerName="init" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.342230 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f406c3b-dbb1-421e-9cee-334414d6089f" containerName="init" Nov 29 00:19:14 crc kubenswrapper[4931]: E1129 00:19:14.342249 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a72a028-d521-466b-9924-eba4595e3750" containerName="neutron-db-sync" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.342257 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a72a028-d521-466b-9924-eba4595e3750" containerName="neutron-db-sync" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.342466 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f406c3b-dbb1-421e-9cee-334414d6089f" containerName="dnsmasq-dns" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.342494 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a72a028-d521-466b-9924-eba4595e3750" containerName="neutron-db-sync" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.347062 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.363264 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6c948c7-rjvnp"] Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.447409 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm6dg\" (UniqueName: \"kubernetes.io/projected/e845944b-9a05-4246-8613-ce413b638e7d-kube-api-access-nm6dg\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.447593 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-ovsdbserver-sb\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.447647 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-dns-svc\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.447705 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-ovsdbserver-nb\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.447764 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-config\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.511793 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b8d6898dd-wpzld"] Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.513569 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.516926 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.517153 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.517322 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jntj9" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.517602 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.550112 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-dns-svc\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.550192 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-ovsdbserver-nb\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.550298 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-config\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.550375 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm6dg\" (UniqueName: \"kubernetes.io/projected/e845944b-9a05-4246-8613-ce413b638e7d-kube-api-access-nm6dg\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.550431 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-ovsdbserver-sb\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.551672 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-dns-svc\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.551882 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-ovsdbserver-nb\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.552067 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-ovsdbserver-sb\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.552839 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-config\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.566989 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b8d6898dd-wpzld"] Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.589020 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm6dg\" (UniqueName: \"kubernetes.io/projected/e845944b-9a05-4246-8613-ce413b638e7d-kube-api-access-nm6dg\") pod \"dnsmasq-dns-b6c948c7-rjvnp\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.652861 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sqww\" (UniqueName: \"kubernetes.io/projected/f7044c78-4dc6-4934-8d91-e8735d23be54-kube-api-access-2sqww\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.652926 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-ovndb-tls-certs\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.653238 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-config\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.653362 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-combined-ca-bundle\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.653405 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-httpd-config\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.655698 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.670136 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.761587 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-scripts\") pod \"af2b8a70-d017-445e-a7bd-f4b19309abf7\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.761950 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af2b8a70-d017-445e-a7bd-f4b19309abf7-logs\") pod \"af2b8a70-d017-445e-a7bd-f4b19309abf7\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.762067 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-combined-ca-bundle\") pod \"af2b8a70-d017-445e-a7bd-f4b19309abf7\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.762149 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af2b8a70-d017-445e-a7bd-f4b19309abf7-httpd-run\") pod \"af2b8a70-d017-445e-a7bd-f4b19309abf7\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.762224 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dln4c\" (UniqueName: \"kubernetes.io/projected/af2b8a70-d017-445e-a7bd-f4b19309abf7-kube-api-access-dln4c\") pod \"af2b8a70-d017-445e-a7bd-f4b19309abf7\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.762285 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-config-data\") pod \"af2b8a70-d017-445e-a7bd-f4b19309abf7\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.762342 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"af2b8a70-d017-445e-a7bd-f4b19309abf7\" (UID: \"af2b8a70-d017-445e-a7bd-f4b19309abf7\") " Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.762831 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sqww\" (UniqueName: \"kubernetes.io/projected/f7044c78-4dc6-4934-8d91-e8735d23be54-kube-api-access-2sqww\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.762901 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-ovndb-tls-certs\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.762997 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-config\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.763031 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af2b8a70-d017-445e-a7bd-f4b19309abf7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "af2b8a70-d017-445e-a7bd-f4b19309abf7" (UID: "af2b8a70-d017-445e-a7bd-f4b19309abf7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.763060 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-combined-ca-bundle\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.763105 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-httpd-config\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.763188 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af2b8a70-d017-445e-a7bd-f4b19309abf7-logs" (OuterVolumeSpecName: "logs") pod "af2b8a70-d017-445e-a7bd-f4b19309abf7" (UID: "af2b8a70-d017-445e-a7bd-f4b19309abf7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.763211 4931 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/af2b8a70-d017-445e-a7bd-f4b19309abf7-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.774129 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-scripts" (OuterVolumeSpecName: "scripts") pod "af2b8a70-d017-445e-a7bd-f4b19309abf7" (UID: "af2b8a70-d017-445e-a7bd-f4b19309abf7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.779014 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-config\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.779537 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-combined-ca-bundle\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.781198 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af2b8a70-d017-445e-a7bd-f4b19309abf7-kube-api-access-dln4c" (OuterVolumeSpecName: "kube-api-access-dln4c") pod "af2b8a70-d017-445e-a7bd-f4b19309abf7" (UID: "af2b8a70-d017-445e-a7bd-f4b19309abf7"). InnerVolumeSpecName "kube-api-access-dln4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.788710 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "af2b8a70-d017-445e-a7bd-f4b19309abf7" (UID: "af2b8a70-d017-445e-a7bd-f4b19309abf7"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.797650 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-httpd-config\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.799286 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sqww\" (UniqueName: \"kubernetes.io/projected/f7044c78-4dc6-4934-8d91-e8735d23be54-kube-api-access-2sqww\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.815314 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-ovndb-tls-certs\") pod \"neutron-b8d6898dd-wpzld\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.834264 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af2b8a70-d017-445e-a7bd-f4b19309abf7" (UID: "af2b8a70-d017-445e-a7bd-f4b19309abf7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.862328 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.865967 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.865992 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af2b8a70-d017-445e-a7bd-f4b19309abf7-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.866000 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.866010 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dln4c\" (UniqueName: \"kubernetes.io/projected/af2b8a70-d017-445e-a7bd-f4b19309abf7-kube-api-access-dln4c\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.866031 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.904260 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.969051 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.979056 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-config-data" (OuterVolumeSpecName: "config-data") pod "af2b8a70-d017-445e-a7bd-f4b19309abf7" (UID: "af2b8a70-d017-445e-a7bd-f4b19309abf7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.980550 4931 generic.go:334] "Generic (PLEG): container finished" podID="af2b8a70-d017-445e-a7bd-f4b19309abf7" containerID="7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd" exitCode=0 Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.980590 4931 generic.go:334] "Generic (PLEG): container finished" podID="af2b8a70-d017-445e-a7bd-f4b19309abf7" containerID="99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100" exitCode=143 Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.980787 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.984009 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"af2b8a70-d017-445e-a7bd-f4b19309abf7","Type":"ContainerDied","Data":"7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd"} Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.984055 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"af2b8a70-d017-445e-a7bd-f4b19309abf7","Type":"ContainerDied","Data":"99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100"} Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.984071 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"af2b8a70-d017-445e-a7bd-f4b19309abf7","Type":"ContainerDied","Data":"43897677b6508b47fb745cd281b673a69571c5af302ce9c553172b7cbeb86d1b"} Nov 29 00:19:14 crc kubenswrapper[4931]: I1129 00:19:14.984087 4931 scope.go:117] "RemoveContainer" containerID="7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.010549 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dn48v" event={"ID":"02ce26c3-6755-479e-93c4-295ef8a3ba2a","Type":"ContainerStarted","Data":"3eb05edd39e3fe12742b69fd09c8953dee41262283efea74b3bd4418cca0f085"} Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.013138 4931 generic.go:334] "Generic (PLEG): container finished" podID="9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" containerID="53fece3204c7e5ba3c8b6b648c62a2dd8b6e51738b70debd3cfdcf677bd17b2f" exitCode=0 Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.013162 4931 generic.go:334] "Generic (PLEG): container finished" podID="9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" containerID="351b54a1ca62d1073c7e1968feda227f49aa0edeb9d4376f3ea4419fb1c96403" exitCode=143 Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.013228 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c","Type":"ContainerDied","Data":"53fece3204c7e5ba3c8b6b648c62a2dd8b6e51738b70debd3cfdcf677bd17b2f"} Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.013252 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c","Type":"ContainerDied","Data":"351b54a1ca62d1073c7e1968feda227f49aa0edeb9d4376f3ea4419fb1c96403"} Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.039221 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-dn48v" podStartSLOduration=3.149300339 podStartE2EDuration="43.03920341s" podCreationTimestamp="2025-11-29 00:18:32 +0000 UTC" firstStartedPulling="2025-11-29 00:18:33.600847872 +0000 UTC m=+1126.762741104" lastFinishedPulling="2025-11-29 00:19:13.490750943 +0000 UTC m=+1166.652644175" observedRunningTime="2025-11-29 00:19:15.029206333 +0000 UTC m=+1168.191099585" watchObservedRunningTime="2025-11-29 00:19:15.03920341 +0000 UTC m=+1168.201096642" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.057982 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.071196 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af2b8a70-d017-445e-a7bd-f4b19309abf7-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.080971 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.098424 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:19:15 crc kubenswrapper[4931]: E1129 00:19:15.098814 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af2b8a70-d017-445e-a7bd-f4b19309abf7" containerName="glance-httpd" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.098841 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="af2b8a70-d017-445e-a7bd-f4b19309abf7" containerName="glance-httpd" Nov 29 00:19:15 crc kubenswrapper[4931]: E1129 00:19:15.098872 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af2b8a70-d017-445e-a7bd-f4b19309abf7" containerName="glance-log" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.098880 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="af2b8a70-d017-445e-a7bd-f4b19309abf7" containerName="glance-log" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.099450 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="af2b8a70-d017-445e-a7bd-f4b19309abf7" containerName="glance-httpd" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.099475 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="af2b8a70-d017-445e-a7bd-f4b19309abf7" containerName="glance-log" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.108299 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.112095 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.112465 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.112544 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.160280 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.228655 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af2b8a70-d017-445e-a7bd-f4b19309abf7" path="/var/lib/kubelet/pods/af2b8a70-d017-445e-a7bd-f4b19309abf7/volumes" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.278211 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.278271 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-config-data\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.278307 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95707751-9dd6-4fd3-a334-b9fafcccbe11-logs\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.278343 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95707751-9dd6-4fd3-a334-b9fafcccbe11-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.278433 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.278536 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-scripts\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.278563 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkgxz\" (UniqueName: \"kubernetes.io/projected/95707751-9dd6-4fd3-a334-b9fafcccbe11-kube-api-access-bkgxz\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.278588 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.380582 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.380683 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-scripts\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.380731 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkgxz\" (UniqueName: \"kubernetes.io/projected/95707751-9dd6-4fd3-a334-b9fafcccbe11-kube-api-access-bkgxz\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.380758 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.380815 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.380849 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-config-data\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.380893 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95707751-9dd6-4fd3-a334-b9fafcccbe11-logs\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.380914 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95707751-9dd6-4fd3-a334-b9fafcccbe11-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.382159 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.384136 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95707751-9dd6-4fd3-a334-b9fafcccbe11-logs\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.396850 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95707751-9dd6-4fd3-a334-b9fafcccbe11-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.399465 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-config-data\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.400202 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-scripts\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.402256 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.403614 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.407389 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkgxz\" (UniqueName: \"kubernetes.io/projected/95707751-9dd6-4fd3-a334-b9fafcccbe11-kube-api-access-bkgxz\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.413122 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6c948c7-rjvnp"] Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.432401 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.452380 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.684091 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b8d6898dd-wpzld"] Nov 29 00:19:15 crc kubenswrapper[4931]: I1129 00:19:15.807502 4931 scope.go:117] "RemoveContainer" containerID="99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100" Nov 29 00:19:15 crc kubenswrapper[4931]: W1129 00:19:15.836497 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7044c78_4dc6_4934_8d91_e8735d23be54.slice/crio-5e8a0a30fa98196d903e4530a4314cdeb7a7c0846a39ed5a04d32e450b610830 WatchSource:0}: Error finding container 5e8a0a30fa98196d903e4530a4314cdeb7a7c0846a39ed5a04d32e450b610830: Status 404 returned error can't find the container with id 5e8a0a30fa98196d903e4530a4314cdeb7a7c0846a39ed5a04d32e450b610830 Nov 29 00:19:15 crc kubenswrapper[4931]: W1129 00:19:15.841150 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode845944b_9a05_4246_8613_ce413b638e7d.slice/crio-c2caa3e407d3792926852a99a95cf7804a335bfdbfcfb6fb1c434eb8cb45592c WatchSource:0}: Error finding container c2caa3e407d3792926852a99a95cf7804a335bfdbfcfb6fb1c434eb8cb45592c: Status 404 returned error can't find the container with id c2caa3e407d3792926852a99a95cf7804a335bfdbfcfb6fb1c434eb8cb45592c Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.064899 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.065481 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" event={"ID":"e845944b-9a05-4246-8613-ce413b638e7d","Type":"ContainerStarted","Data":"c2caa3e407d3792926852a99a95cf7804a335bfdbfcfb6fb1c434eb8cb45592c"} Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.066588 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b8d6898dd-wpzld" event={"ID":"f7044c78-4dc6-4934-8d91-e8735d23be54","Type":"ContainerStarted","Data":"5e8a0a30fa98196d903e4530a4314cdeb7a7c0846a39ed5a04d32e450b610830"} Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.075752 4931 scope.go:117] "RemoveContainer" containerID="7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd" Nov 29 00:19:16 crc kubenswrapper[4931]: E1129 00:19:16.079361 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd\": container with ID starting with 7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd not found: ID does not exist" containerID="7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.079601 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd"} err="failed to get container status \"7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd\": rpc error: code = NotFound desc = could not find container \"7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd\": container with ID starting with 7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd not found: ID does not exist" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.079632 4931 scope.go:117] "RemoveContainer" containerID="99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.079890 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c","Type":"ContainerDied","Data":"c383c40c38fccee4291233801ec9208211f95197ad31cde74989ff9feba92b01"} Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.079960 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" podUID="d3c1e52d-f484-4f0c-abbd-2f27a7e83592" containerName="dnsmasq-dns" containerID="cri-o://813b56703bc6e2cfd71c7ba156b1834f10a2c64959025602085c975a22a40b42" gracePeriod=10 Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.079984 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: E1129 00:19:16.080303 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100\": container with ID starting with 99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100 not found: ID does not exist" containerID="99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.080336 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100"} err="failed to get container status \"99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100\": rpc error: code = NotFound desc = could not find container \"99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100\": container with ID starting with 99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100 not found: ID does not exist" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.080358 4931 scope.go:117] "RemoveContainer" containerID="7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.080816 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd"} err="failed to get container status \"7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd\": rpc error: code = NotFound desc = could not find container \"7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd\": container with ID starting with 7a952446cc4a2d8bde9a0b11149790b1b6485fc0e8af3b27bc1ac769795cf5bd not found: ID does not exist" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.080853 4931 scope.go:117] "RemoveContainer" containerID="99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.081607 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100"} err="failed to get container status \"99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100\": rpc error: code = NotFound desc = could not find container \"99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100\": container with ID starting with 99d7f671cea73a29d2e56c2006cb07e7c2ad996e5016f208256b6b8967f87100 not found: ID does not exist" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.081631 4931 scope.go:117] "RemoveContainer" containerID="53fece3204c7e5ba3c8b6b648c62a2dd8b6e51738b70debd3cfdcf677bd17b2f" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.158074 4931 scope.go:117] "RemoveContainer" containerID="351b54a1ca62d1073c7e1968feda227f49aa0edeb9d4376f3ea4419fb1c96403" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.200395 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-config-data\") pod \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.200513 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-combined-ca-bundle\") pod \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.200547 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-scripts\") pod \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.200567 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-logs\") pod \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.200588 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-httpd-run\") pod \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.200620 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptbb2\" (UniqueName: \"kubernetes.io/projected/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-kube-api-access-ptbb2\") pod \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.200757 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\" (UID: \"9b36d110-9dcc-4434-8b56-e84e1ee7ab8c\") " Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.206038 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" (UID: "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.210651 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-logs" (OuterVolumeSpecName: "logs") pod "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" (UID: "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.210953 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" (UID: "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.228994 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-scripts" (OuterVolumeSpecName: "scripts") pod "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" (UID: "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.230119 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-kube-api-access-ptbb2" (OuterVolumeSpecName: "kube-api-access-ptbb2") pod "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" (UID: "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c"). InnerVolumeSpecName "kube-api-access-ptbb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.251718 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" (UID: "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.302531 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.302566 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.302751 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.302758 4931 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.302767 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptbb2\" (UniqueName: \"kubernetes.io/projected/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-kube-api-access-ptbb2\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.302789 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.348291 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-config-data" (OuterVolumeSpecName: "config-data") pod "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" (UID: "9b36d110-9dcc-4434-8b56-e84e1ee7ab8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.355533 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.418571 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.418903 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.466890 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.476409 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.527368 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:19:16 crc kubenswrapper[4931]: E1129 00:19:16.527814 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" containerName="glance-httpd" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.527852 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" containerName="glance-httpd" Nov 29 00:19:16 crc kubenswrapper[4931]: E1129 00:19:16.527873 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" containerName="glance-log" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.527881 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" containerName="glance-log" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.528109 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" containerName="glance-httpd" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.528136 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" containerName="glance-log" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.530732 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.535446 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.538980 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.541240 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.623586 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.623644 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.623698 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.623762 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfqg5\" (UniqueName: \"kubernetes.io/projected/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-kube-api-access-mfqg5\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.623798 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.623876 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.623919 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-logs\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.623943 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.735342 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.735421 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfqg5\" (UniqueName: \"kubernetes.io/projected/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-kube-api-access-mfqg5\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.735451 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.735498 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.735534 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-logs\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.735551 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.735588 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.735607 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.736064 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.741235 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.746060 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-logs\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.756594 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.756715 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.776825 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.776883 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.791094 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.821426 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfqg5\" (UniqueName: \"kubernetes.io/projected/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-kube-api-access-mfqg5\") pod \"glance-default-internal-api-0\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.864359 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:16 crc kubenswrapper[4931]: I1129 00:19:16.957615 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.108054 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95707751-9dd6-4fd3-a334-b9fafcccbe11","Type":"ContainerStarted","Data":"46fed1a5e9849dc21241f2a62f29996cf8bcb8d5d11ada3effd0764846a9ceb8"} Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.254506 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b36d110-9dcc-4434-8b56-e84e1ee7ab8c" path="/var/lib/kubelet/pods/9b36d110-9dcc-4434-8b56-e84e1ee7ab8c/volumes" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.425457 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.540964 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-bcd7d89b9-l4sjs"] Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.557226 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bcd7d89b9-l4sjs"] Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.557322 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.560319 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.560535 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.657206 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-combined-ca-bundle\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.657257 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-public-tls-certs\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.657444 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4vn8\" (UniqueName: \"kubernetes.io/projected/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-kube-api-access-p4vn8\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.657507 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-config\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.657535 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-internal-tls-certs\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.657590 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-httpd-config\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.657653 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-ovndb-tls-certs\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.759008 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-internal-tls-certs\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.759109 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-httpd-config\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.759149 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-ovndb-tls-certs\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.759182 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-combined-ca-bundle\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.759217 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-public-tls-certs\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.759413 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4vn8\" (UniqueName: \"kubernetes.io/projected/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-kube-api-access-p4vn8\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.759478 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-config\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.764546 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-internal-tls-certs\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.766931 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-combined-ca-bundle\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.768684 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-ovndb-tls-certs\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.769231 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-config\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.769517 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-httpd-config\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.771323 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-public-tls-certs\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.802204 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4vn8\" (UniqueName: \"kubernetes.io/projected/e1ff7fdf-846a-4772-b8d8-fee7b15eed58-kube-api-access-p4vn8\") pod \"neutron-bcd7d89b9-l4sjs\" (UID: \"e1ff7fdf-846a-4772-b8d8-fee7b15eed58\") " pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:17 crc kubenswrapper[4931]: I1129 00:19:17.903025 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:18 crc kubenswrapper[4931]: I1129 00:19:18.130571 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1","Type":"ContainerStarted","Data":"c27a49a679d4578f60ec9719f2f389e5bb4cfa24fb3a156b4e174b91a8ff51d0"} Nov 29 00:19:18 crc kubenswrapper[4931]: I1129 00:19:18.533962 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bcd7d89b9-l4sjs"] Nov 29 00:19:18 crc kubenswrapper[4931]: W1129 00:19:18.546245 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1ff7fdf_846a_4772_b8d8_fee7b15eed58.slice/crio-1eed6cb995c2cc5ec3450562ce27de196f0cc1150c861b2055cda0ab3385caa6 WatchSource:0}: Error finding container 1eed6cb995c2cc5ec3450562ce27de196f0cc1150c861b2055cda0ab3385caa6: Status 404 returned error can't find the container with id 1eed6cb995c2cc5ec3450562ce27de196f0cc1150c861b2055cda0ab3385caa6 Nov 29 00:19:19 crc kubenswrapper[4931]: I1129 00:19:19.145308 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bcd7d89b9-l4sjs" event={"ID":"e1ff7fdf-846a-4772-b8d8-fee7b15eed58","Type":"ContainerStarted","Data":"1eed6cb995c2cc5ec3450562ce27de196f0cc1150c861b2055cda0ab3385caa6"} Nov 29 00:19:20 crc kubenswrapper[4931]: I1129 00:19:20.156294 4931 generic.go:334] "Generic (PLEG): container finished" podID="d3c1e52d-f484-4f0c-abbd-2f27a7e83592" containerID="813b56703bc6e2cfd71c7ba156b1834f10a2c64959025602085c975a22a40b42" exitCode=0 Nov 29 00:19:20 crc kubenswrapper[4931]: I1129 00:19:20.156379 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" event={"ID":"d3c1e52d-f484-4f0c-abbd-2f27a7e83592","Type":"ContainerDied","Data":"813b56703bc6e2cfd71c7ba156b1834f10a2c64959025602085c975a22a40b42"} Nov 29 00:19:20 crc kubenswrapper[4931]: I1129 00:19:20.157757 4931 generic.go:334] "Generic (PLEG): container finished" podID="8e6ded52-583f-44e7-853c-30c354886ae9" containerID="f1ad50d40ba416d98d59313fc595dc27e2284b8f1c20fc2459e688f103bfe16d" exitCode=0 Nov 29 00:19:20 crc kubenswrapper[4931]: I1129 00:19:20.157792 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wknnq" event={"ID":"8e6ded52-583f-44e7-853c-30c354886ae9","Type":"ContainerDied","Data":"f1ad50d40ba416d98d59313fc595dc27e2284b8f1c20fc2459e688f103bfe16d"} Nov 29 00:19:20 crc kubenswrapper[4931]: I1129 00:19:20.161081 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" event={"ID":"e845944b-9a05-4246-8613-ce413b638e7d","Type":"ContainerStarted","Data":"1068e27541527caaacb8cf808b4362cf4d0d8313d1833865665b72eb246618c7"} Nov 29 00:19:20 crc kubenswrapper[4931]: I1129 00:19:20.163953 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b8d6898dd-wpzld" event={"ID":"f7044c78-4dc6-4934-8d91-e8735d23be54","Type":"ContainerStarted","Data":"4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3"} Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.176674 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1","Type":"ContainerStarted","Data":"3ab41f0a353db4ad8930f11915ecf91f2f4d053f82ad4c8a9cde1a3e0955ed6f"} Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.179410 4931 generic.go:334] "Generic (PLEG): container finished" podID="e845944b-9a05-4246-8613-ce413b638e7d" containerID="1068e27541527caaacb8cf808b4362cf4d0d8313d1833865665b72eb246618c7" exitCode=0 Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.179452 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" event={"ID":"e845944b-9a05-4246-8613-ce413b638e7d","Type":"ContainerDied","Data":"1068e27541527caaacb8cf808b4362cf4d0d8313d1833865665b72eb246618c7"} Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.182514 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95707751-9dd6-4fd3-a334-b9fafcccbe11","Type":"ContainerStarted","Data":"a8fda88fd35b023ddcfa5bdd32d75ccc4abad9b5fb0dffa414b56d66ba10dd4d"} Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.188294 4931 generic.go:334] "Generic (PLEG): container finished" podID="6041acf3-4620-48a5-8455-6c8cdb50cb0b" containerID="691315ec4d054a167417bbadfd4ddb939e88c1eed51bf22e2853223ea252e573" exitCode=0 Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.188346 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z2cs6" event={"ID":"6041acf3-4620-48a5-8455-6c8cdb50cb0b","Type":"ContainerDied","Data":"691315ec4d054a167417bbadfd4ddb939e88c1eed51bf22e2853223ea252e573"} Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.191173 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bcd7d89b9-l4sjs" event={"ID":"e1ff7fdf-846a-4772-b8d8-fee7b15eed58","Type":"ContainerStarted","Data":"b824c8ba6487ea94b4edb439991991adfbb5ac86d2ffa88f94cc5a5a68eefe14"} Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.343377 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.343772 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.366080 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-765c4d856b-6tkcn" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.459264 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.459301 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:19:21 crc kubenswrapper[4931]: I1129 00:19:21.467095 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-79d99cf788-wwqzd" podUID="4003b615-a779-45a5-b95b-913c80226f4a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.753742 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.759000 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z2cs6" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.766997 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888517 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-scripts\") pod \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888571 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-ovsdbserver-sb\") pod \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888606 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-credential-keys\") pod \"8e6ded52-583f-44e7-853c-30c354886ae9\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888630 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l42th\" (UniqueName: \"kubernetes.io/projected/6041acf3-4620-48a5-8455-6c8cdb50cb0b-kube-api-access-l42th\") pod \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888651 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-scripts\") pod \"8e6ded52-583f-44e7-853c-30c354886ae9\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888672 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-config-data\") pod \"8e6ded52-583f-44e7-853c-30c354886ae9\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888698 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt2kz\" (UniqueName: \"kubernetes.io/projected/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-kube-api-access-dt2kz\") pod \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888719 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-fernet-keys\") pod \"8e6ded52-583f-44e7-853c-30c354886ae9\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888745 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-combined-ca-bundle\") pod \"8e6ded52-583f-44e7-853c-30c354886ae9\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888773 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-dns-svc\") pod \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888831 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-ovsdbserver-nb\") pod \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888856 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-config\") pod \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\" (UID: \"d3c1e52d-f484-4f0c-abbd-2f27a7e83592\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888884 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-config-data\") pod \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888908 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6041acf3-4620-48a5-8455-6c8cdb50cb0b-logs\") pod \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888932 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqvf9\" (UniqueName: \"kubernetes.io/projected/8e6ded52-583f-44e7-853c-30c354886ae9-kube-api-access-cqvf9\") pod \"8e6ded52-583f-44e7-853c-30c354886ae9\" (UID: \"8e6ded52-583f-44e7-853c-30c354886ae9\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.888999 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-combined-ca-bundle\") pod \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\" (UID: \"6041acf3-4620-48a5-8455-6c8cdb50cb0b\") " Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.899989 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6041acf3-4620-48a5-8455-6c8cdb50cb0b-logs" (OuterVolumeSpecName: "logs") pod "6041acf3-4620-48a5-8455-6c8cdb50cb0b" (UID: "6041acf3-4620-48a5-8455-6c8cdb50cb0b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.911978 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6041acf3-4620-48a5-8455-6c8cdb50cb0b-kube-api-access-l42th" (OuterVolumeSpecName: "kube-api-access-l42th") pod "6041acf3-4620-48a5-8455-6c8cdb50cb0b" (UID: "6041acf3-4620-48a5-8455-6c8cdb50cb0b"). InnerVolumeSpecName "kube-api-access-l42th". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.912074 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8e6ded52-583f-44e7-853c-30c354886ae9" (UID: "8e6ded52-583f-44e7-853c-30c354886ae9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.917984 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-scripts" (OuterVolumeSpecName: "scripts") pod "6041acf3-4620-48a5-8455-6c8cdb50cb0b" (UID: "6041acf3-4620-48a5-8455-6c8cdb50cb0b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.931003 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-kube-api-access-dt2kz" (OuterVolumeSpecName: "kube-api-access-dt2kz") pod "d3c1e52d-f484-4f0c-abbd-2f27a7e83592" (UID: "d3c1e52d-f484-4f0c-abbd-2f27a7e83592"). InnerVolumeSpecName "kube-api-access-dt2kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.951399 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-scripts" (OuterVolumeSpecName: "scripts") pod "8e6ded52-583f-44e7-853c-30c354886ae9" (UID: "8e6ded52-583f-44e7-853c-30c354886ae9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.966168 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8e6ded52-583f-44e7-853c-30c354886ae9" (UID: "8e6ded52-583f-44e7-853c-30c354886ae9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.966300 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e6ded52-583f-44e7-853c-30c354886ae9-kube-api-access-cqvf9" (OuterVolumeSpecName: "kube-api-access-cqvf9") pod "8e6ded52-583f-44e7-853c-30c354886ae9" (UID: "8e6ded52-583f-44e7-853c-30c354886ae9"). InnerVolumeSpecName "kube-api-access-cqvf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.976353 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e6ded52-583f-44e7-853c-30c354886ae9" (UID: "8e6ded52-583f-44e7-853c-30c354886ae9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.986449 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6041acf3-4620-48a5-8455-6c8cdb50cb0b" (UID: "6041acf3-4620-48a5-8455-6c8cdb50cb0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.993013 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6041acf3-4620-48a5-8455-6c8cdb50cb0b-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.993048 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqvf9\" (UniqueName: \"kubernetes.io/projected/8e6ded52-583f-44e7-853c-30c354886ae9-kube-api-access-cqvf9\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.993060 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.993069 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.993077 4931 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.993086 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l42th\" (UniqueName: \"kubernetes.io/projected/6041acf3-4620-48a5-8455-6c8cdb50cb0b-kube-api-access-l42th\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.993093 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.993102 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt2kz\" (UniqueName: \"kubernetes.io/projected/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-kube-api-access-dt2kz\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.993109 4931 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:22 crc kubenswrapper[4931]: I1129 00:19:22.993117 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.018387 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-config-data" (OuterVolumeSpecName: "config-data") pod "6041acf3-4620-48a5-8455-6c8cdb50cb0b" (UID: "6041acf3-4620-48a5-8455-6c8cdb50cb0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.024278 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-config-data" (OuterVolumeSpecName: "config-data") pod "8e6ded52-583f-44e7-853c-30c354886ae9" (UID: "8e6ded52-583f-44e7-853c-30c354886ae9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.038977 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-config" (OuterVolumeSpecName: "config") pod "d3c1e52d-f484-4f0c-abbd-2f27a7e83592" (UID: "d3c1e52d-f484-4f0c-abbd-2f27a7e83592"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.081795 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d3c1e52d-f484-4f0c-abbd-2f27a7e83592" (UID: "d3c1e52d-f484-4f0c-abbd-2f27a7e83592"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.086328 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d3c1e52d-f484-4f0c-abbd-2f27a7e83592" (UID: "d3c1e52d-f484-4f0c-abbd-2f27a7e83592"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.094785 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.094832 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e6ded52-583f-44e7-853c-30c354886ae9-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.094843 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.094854 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.094863 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6041acf3-4620-48a5-8455-6c8cdb50cb0b-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.099607 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d3c1e52d-f484-4f0c-abbd-2f27a7e83592" (UID: "d3c1e52d-f484-4f0c-abbd-2f27a7e83592"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.196996 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3c1e52d-f484-4f0c-abbd-2f27a7e83592-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.239614 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z2cs6" event={"ID":"6041acf3-4620-48a5-8455-6c8cdb50cb0b","Type":"ContainerDied","Data":"9c1a6439c2697030aad62ca0f75ba2cd5d9e6bb1987b4c6a75e193c34da587d3"} Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.239651 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c1a6439c2697030aad62ca0f75ba2cd5d9e6bb1987b4c6a75e193c34da587d3" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.239716 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z2cs6" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.247525 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" event={"ID":"d3c1e52d-f484-4f0c-abbd-2f27a7e83592","Type":"ContainerDied","Data":"75089c71cb2000d0679a5069803060ed7da1e8910e5654f0e54f26af6da33e9d"} Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.247676 4931 scope.go:117] "RemoveContainer" containerID="813b56703bc6e2cfd71c7ba156b1834f10a2c64959025602085c975a22a40b42" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.247886 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.265560 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wknnq" event={"ID":"8e6ded52-583f-44e7-853c-30c354886ae9","Type":"ContainerDied","Data":"6634eae225d37b06dc6f94e01633c6409d554b410423aaaa9acf665f68a4f6fc"} Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.265931 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6634eae225d37b06dc6f94e01633c6409d554b410423aaaa9acf665f68a4f6fc" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.265994 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wknnq" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.288903 4931 generic.go:334] "Generic (PLEG): container finished" podID="27762870-383b-4e1b-8d28-e6e7c96d635a" containerID="9a39ab75b839a4c6e4d50a13762239245f3d69a686155574f10f887436ee76e3" exitCode=0 Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.289153 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pml9v" event={"ID":"27762870-383b-4e1b-8d28-e6e7c96d635a","Type":"ContainerDied","Data":"9a39ab75b839a4c6e4d50a13762239245f3d69a686155574f10f887436ee76e3"} Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.360245 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6b5869645b-zlntx"] Nov 29 00:19:23 crc kubenswrapper[4931]: E1129 00:19:23.362235 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6041acf3-4620-48a5-8455-6c8cdb50cb0b" containerName="placement-db-sync" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.362348 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="6041acf3-4620-48a5-8455-6c8cdb50cb0b" containerName="placement-db-sync" Nov 29 00:19:23 crc kubenswrapper[4931]: E1129 00:19:23.362466 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6ded52-583f-44e7-853c-30c354886ae9" containerName="keystone-bootstrap" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.362532 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6ded52-583f-44e7-853c-30c354886ae9" containerName="keystone-bootstrap" Nov 29 00:19:23 crc kubenswrapper[4931]: E1129 00:19:23.362592 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3c1e52d-f484-4f0c-abbd-2f27a7e83592" containerName="init" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.362638 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3c1e52d-f484-4f0c-abbd-2f27a7e83592" containerName="init" Nov 29 00:19:23 crc kubenswrapper[4931]: E1129 00:19:23.362703 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3c1e52d-f484-4f0c-abbd-2f27a7e83592" containerName="dnsmasq-dns" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.362753 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3c1e52d-f484-4f0c-abbd-2f27a7e83592" containerName="dnsmasq-dns" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.363050 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e6ded52-583f-44e7-853c-30c354886ae9" containerName="keystone-bootstrap" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.378958 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3c1e52d-f484-4f0c-abbd-2f27a7e83592" containerName="dnsmasq-dns" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.379213 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="6041acf3-4620-48a5-8455-6c8cdb50cb0b" containerName="placement-db-sync" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.380704 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.385804 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.387386 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xdxfw" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.387530 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.387628 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.393775 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.405006 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6b5869645b-zlntx"] Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.430902 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56798b757f-jxjxc"] Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.436939 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56798b757f-jxjxc"] Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.505944 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-public-tls-certs\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.505997 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-config-data\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.506016 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48538810-7ec7-4d21-9cf3-1c0ee151eee5-logs\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.506061 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-internal-tls-certs\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.506874 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-scripts\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.506925 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-combined-ca-bundle\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.507268 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc95x\" (UniqueName: \"kubernetes.io/projected/48538810-7ec7-4d21-9cf3-1c0ee151eee5-kube-api-access-bc95x\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.609425 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc95x\" (UniqueName: \"kubernetes.io/projected/48538810-7ec7-4d21-9cf3-1c0ee151eee5-kube-api-access-bc95x\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.609517 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-public-tls-certs\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.609539 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-config-data\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.609555 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48538810-7ec7-4d21-9cf3-1c0ee151eee5-logs\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.609604 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-internal-tls-certs\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.609636 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-scripts\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.609654 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-combined-ca-bundle\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.610464 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48538810-7ec7-4d21-9cf3-1c0ee151eee5-logs\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.613659 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-scripts\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.614154 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-combined-ca-bundle\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.616664 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-public-tls-certs\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.617950 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-config-data\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.625702 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48538810-7ec7-4d21-9cf3-1c0ee151eee5-internal-tls-certs\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.626461 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc95x\" (UniqueName: \"kubernetes.io/projected/48538810-7ec7-4d21-9cf3-1c0ee151eee5-kube-api-access-bc95x\") pod \"placement-6b5869645b-zlntx\" (UID: \"48538810-7ec7-4d21-9cf3-1c0ee151eee5\") " pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.716884 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.918630 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-84d76c555c-shgd6"] Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.921007 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.924351 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.924350 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.924707 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bwpgd" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.924777 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.924810 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.930574 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 29 00:19:23 crc kubenswrapper[4931]: I1129 00:19:23.934286 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-84d76c555c-shgd6"] Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.121987 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-fernet-keys\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.122032 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92z8b\" (UniqueName: \"kubernetes.io/projected/aff73224-6284-4ef1-ace3-06eb11e904b8-kube-api-access-92z8b\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.122621 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-combined-ca-bundle\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.122693 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-config-data\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.122710 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-internal-tls-certs\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.122743 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-public-tls-certs\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.122768 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-credential-keys\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.122906 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-scripts\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.224622 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-fernet-keys\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.225239 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92z8b\" (UniqueName: \"kubernetes.io/projected/aff73224-6284-4ef1-ace3-06eb11e904b8-kube-api-access-92z8b\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.225699 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-combined-ca-bundle\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.225836 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-config-data\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.225922 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-internal-tls-certs\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.226026 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-public-tls-certs\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.226115 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-credential-keys\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.226285 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-scripts\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.234092 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-scripts\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.235922 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-combined-ca-bundle\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.242999 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-fernet-keys\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.243345 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-config-data\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.245728 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-internal-tls-certs\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.246252 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-credential-keys\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.247617 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92z8b\" (UniqueName: \"kubernetes.io/projected/aff73224-6284-4ef1-ace3-06eb11e904b8-kube-api-access-92z8b\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.255602 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aff73224-6284-4ef1-ace3-06eb11e904b8-public-tls-certs\") pod \"keystone-84d76c555c-shgd6\" (UID: \"aff73224-6284-4ef1-ace3-06eb11e904b8\") " pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:24 crc kubenswrapper[4931]: I1129 00:19:24.547525 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:25 crc kubenswrapper[4931]: I1129 00:19:25.222563 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3c1e52d-f484-4f0c-abbd-2f27a7e83592" path="/var/lib/kubelet/pods/d3c1e52d-f484-4f0c-abbd-2f27a7e83592/volumes" Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.055696 4931 scope.go:117] "RemoveContainer" containerID="91982a165b703336f2c42a80fccbcc2bd30cf39c7246caed00c275a969f81a58" Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.204056 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pml9v" Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.370106 4931 generic.go:334] "Generic (PLEG): container finished" podID="02ce26c3-6755-479e-93c4-295ef8a3ba2a" containerID="3eb05edd39e3fe12742b69fd09c8953dee41262283efea74b3bd4418cca0f085" exitCode=0 Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.370276 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dn48v" event={"ID":"02ce26c3-6755-479e-93c4-295ef8a3ba2a","Type":"ContainerDied","Data":"3eb05edd39e3fe12742b69fd09c8953dee41262283efea74b3bd4418cca0f085"} Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.370773 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsl75\" (UniqueName: \"kubernetes.io/projected/27762870-383b-4e1b-8d28-e6e7c96d635a-kube-api-access-dsl75\") pod \"27762870-383b-4e1b-8d28-e6e7c96d635a\" (UID: \"27762870-383b-4e1b-8d28-e6e7c96d635a\") " Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.372204 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27762870-383b-4e1b-8d28-e6e7c96d635a-combined-ca-bundle\") pod \"27762870-383b-4e1b-8d28-e6e7c96d635a\" (UID: \"27762870-383b-4e1b-8d28-e6e7c96d635a\") " Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.372254 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/27762870-383b-4e1b-8d28-e6e7c96d635a-db-sync-config-data\") pod \"27762870-383b-4e1b-8d28-e6e7c96d635a\" (UID: \"27762870-383b-4e1b-8d28-e6e7c96d635a\") " Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.405168 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27762870-383b-4e1b-8d28-e6e7c96d635a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "27762870-383b-4e1b-8d28-e6e7c96d635a" (UID: "27762870-383b-4e1b-8d28-e6e7c96d635a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.405375 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27762870-383b-4e1b-8d28-e6e7c96d635a-kube-api-access-dsl75" (OuterVolumeSpecName: "kube-api-access-dsl75") pod "27762870-383b-4e1b-8d28-e6e7c96d635a" (UID: "27762870-383b-4e1b-8d28-e6e7c96d635a"). InnerVolumeSpecName "kube-api-access-dsl75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.405396 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pml9v" event={"ID":"27762870-383b-4e1b-8d28-e6e7c96d635a","Type":"ContainerDied","Data":"bb357b079f0955d44569c9ed85ab2a66f066f72acc2320672b75a559cf0ba818"} Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.405430 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb357b079f0955d44569c9ed85ab2a66f066f72acc2320672b75a559cf0ba818" Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.405483 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pml9v" Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.456946 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27762870-383b-4e1b-8d28-e6e7c96d635a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27762870-383b-4e1b-8d28-e6e7c96d635a" (UID: "27762870-383b-4e1b-8d28-e6e7c96d635a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.474472 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27762870-383b-4e1b-8d28-e6e7c96d635a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.474784 4931 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/27762870-383b-4e1b-8d28-e6e7c96d635a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.474797 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsl75\" (UniqueName: \"kubernetes.io/projected/27762870-383b-4e1b-8d28-e6e7c96d635a-kube-api-access-dsl75\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.635906 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6b5869645b-zlntx"] Nov 29 00:19:26 crc kubenswrapper[4931]: W1129 00:19:26.647964 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48538810_7ec7_4d21_9cf3_1c0ee151eee5.slice/crio-53066ef934f2ac4d1ac8ad07992aa40de20dc752707893224b30d313e73540e9 WatchSource:0}: Error finding container 53066ef934f2ac4d1ac8ad07992aa40de20dc752707893224b30d313e73540e9: Status 404 returned error can't find the container with id 53066ef934f2ac4d1ac8ad07992aa40de20dc752707893224b30d313e73540e9 Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.738167 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-84d76c555c-shgd6"] Nov 29 00:19:26 crc kubenswrapper[4931]: I1129 00:19:26.791166 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-56798b757f-jxjxc" podUID="d3c1e52d-f484-4f0c-abbd-2f27a7e83592" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.145:5353: i/o timeout" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.414935 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1","Type":"ContainerStarted","Data":"33a5411b3d35694169a92a5ac4550140a6c43cf93ab7ceeba00c5f40d6ca9d36"} Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.417015 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-84d76c555c-shgd6" event={"ID":"aff73224-6284-4ef1-ace3-06eb11e904b8","Type":"ContainerStarted","Data":"1857c09a5ee8c6343fc4996f954d6c6b63097634de466345cb1dc13951d12cb0"} Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.417062 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-84d76c555c-shgd6" event={"ID":"aff73224-6284-4ef1-ace3-06eb11e904b8","Type":"ContainerStarted","Data":"c702fda1ff32e8c3f1c4ffcc0cf57be2e15343cac884c7d0afb9fbfe1e070f03"} Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.417110 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.418610 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" event={"ID":"e845944b-9a05-4246-8613-ce413b638e7d","Type":"ContainerStarted","Data":"b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af"} Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.418773 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.420623 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57b51201-4e02-4b29-843d-6bfe84247cd8","Type":"ContainerStarted","Data":"9d01970592fdce0484783d5f757c978cbd0536d5468f72ad79ddae208fd84d3d"} Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.422523 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6b5869645b-zlntx" event={"ID":"48538810-7ec7-4d21-9cf3-1c0ee151eee5","Type":"ContainerStarted","Data":"c3100a045855d95682005fa8e64d345461cee5a427188967de334860fceee891"} Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.422571 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6b5869645b-zlntx" event={"ID":"48538810-7ec7-4d21-9cf3-1c0ee151eee5","Type":"ContainerStarted","Data":"28330dfe42a163f25ad768e9ef9a26dcf067c1e7e241725b4cf0727611719699"} Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.422581 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6b5869645b-zlntx" event={"ID":"48538810-7ec7-4d21-9cf3-1c0ee151eee5","Type":"ContainerStarted","Data":"53066ef934f2ac4d1ac8ad07992aa40de20dc752707893224b30d313e73540e9"} Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.423290 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.423320 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.424924 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95707751-9dd6-4fd3-a334-b9fafcccbe11","Type":"ContainerStarted","Data":"4d5dc2f4eab8c21d8b7a78f5f16f4a198c7245fd51df7958bff094f5455d0450"} Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.426673 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b8d6898dd-wpzld" event={"ID":"f7044c78-4dc6-4934-8d91-e8735d23be54","Type":"ContainerStarted","Data":"7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122"} Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.427202 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.429426 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bcd7d89b9-l4sjs" event={"ID":"e1ff7fdf-846a-4772-b8d8-fee7b15eed58","Type":"ContainerStarted","Data":"b40927572683bfe4b617295dba359cec89d78110e081c88738657ee2456a12d1"} Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.429462 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.503615 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.503595366999999 podStartE2EDuration="11.503595367s" podCreationTimestamp="2025-11-29 00:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:27.489174982 +0000 UTC m=+1180.651068234" watchObservedRunningTime="2025-11-29 00:19:27.503595367 +0000 UTC m=+1180.665488599" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.556261 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-b8d6898dd-wpzld" podStartSLOduration=13.55622804 podStartE2EDuration="13.55622804s" podCreationTimestamp="2025-11-29 00:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:27.554250353 +0000 UTC m=+1180.716143605" watchObservedRunningTime="2025-11-29 00:19:27.55622804 +0000 UTC m=+1180.718121272" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.636194 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-bcd7d89b9-l4sjs" podStartSLOduration=10.636176468 podStartE2EDuration="10.636176468s" podCreationTimestamp="2025-11-29 00:19:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:27.634398647 +0000 UTC m=+1180.796291899" watchObservedRunningTime="2025-11-29 00:19:27.636176468 +0000 UTC m=+1180.798069700" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.680864 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5644b469ff-dnc8z"] Nov 29 00:19:27 crc kubenswrapper[4931]: E1129 00:19:27.681239 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27762870-383b-4e1b-8d28-e6e7c96d635a" containerName="barbican-db-sync" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.681250 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="27762870-383b-4e1b-8d28-e6e7c96d635a" containerName="barbican-db-sync" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.681425 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="27762870-383b-4e1b-8d28-e6e7c96d635a" containerName="barbican-db-sync" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.682269 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.692509 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.692780 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-7wlxl" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.692856 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.695214 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6b65859cc8-zkpcj"] Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.696596 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.707360 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.727913 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5644b469ff-dnc8z"] Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.728480 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=12.728470270999999 podStartE2EDuration="12.728470271s" podCreationTimestamp="2025-11-29 00:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:27.692643081 +0000 UTC m=+1180.854536313" watchObservedRunningTime="2025-11-29 00:19:27.728470271 +0000 UTC m=+1180.890363503" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.777398 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-84d76c555c-shgd6" podStartSLOduration=4.777376937 podStartE2EDuration="4.777376937s" podCreationTimestamp="2025-11-29 00:19:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:27.741180667 +0000 UTC m=+1180.903073909" watchObservedRunningTime="2025-11-29 00:19:27.777376937 +0000 UTC m=+1180.939270169" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.777913 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6b65859cc8-zkpcj"] Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.806095 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6b5869645b-zlntx" podStartSLOduration=4.806077042 podStartE2EDuration="4.806077042s" podCreationTimestamp="2025-11-29 00:19:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:27.798290988 +0000 UTC m=+1180.960184220" watchObservedRunningTime="2025-11-29 00:19:27.806077042 +0000 UTC m=+1180.967970274" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.813195 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-config-data\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.813236 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ba68cea-6add-4e52-96ad-458f708a02e0-logs\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.813261 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-config-data-custom\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.813311 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-combined-ca-bundle\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.813357 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba68cea-6add-4e52-96ad-458f708a02e0-combined-ca-bundle\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.813376 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bvnk\" (UniqueName: \"kubernetes.io/projected/2ba68cea-6add-4e52-96ad-458f708a02e0-kube-api-access-4bvnk\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.813393 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv72j\" (UniqueName: \"kubernetes.io/projected/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-kube-api-access-lv72j\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.813428 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ba68cea-6add-4e52-96ad-458f708a02e0-config-data-custom\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.813461 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-logs\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.813498 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ba68cea-6add-4e52-96ad-458f708a02e0-config-data\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.868581 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" podStartSLOduration=13.868554748 podStartE2EDuration="13.868554748s" podCreationTimestamp="2025-11-29 00:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:27.825213672 +0000 UTC m=+1180.987106914" watchObservedRunningTime="2025-11-29 00:19:27.868554748 +0000 UTC m=+1181.030447980" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.915791 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-combined-ca-bundle\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.920569 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba68cea-6add-4e52-96ad-458f708a02e0-combined-ca-bundle\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.920611 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bvnk\" (UniqueName: \"kubernetes.io/projected/2ba68cea-6add-4e52-96ad-458f708a02e0-kube-api-access-4bvnk\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.920635 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv72j\" (UniqueName: \"kubernetes.io/projected/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-kube-api-access-lv72j\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.920700 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ba68cea-6add-4e52-96ad-458f708a02e0-config-data-custom\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.920773 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-logs\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.920841 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ba68cea-6add-4e52-96ad-458f708a02e0-config-data\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.920890 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-config-data\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.920905 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ba68cea-6add-4e52-96ad-458f708a02e0-logs\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.920932 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-config-data-custom\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.922165 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-logs\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.932326 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ba68cea-6add-4e52-96ad-458f708a02e0-logs\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.950500 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba68cea-6add-4e52-96ad-458f708a02e0-combined-ca-bundle\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.951468 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ba68cea-6add-4e52-96ad-458f708a02e0-config-data\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.962538 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-config-data\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.963536 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv72j\" (UniqueName: \"kubernetes.io/projected/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-kube-api-access-lv72j\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.964108 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6c948c7-rjvnp"] Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.983555 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ba68cea-6add-4e52-96ad-458f708a02e0-config-data-custom\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.985452 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-combined-ca-bundle\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.989875 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8976d4b6b-jvhlw"] Nov 29 00:19:27 crc kubenswrapper[4931]: I1129 00:19:27.991689 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:27.992584 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5941c09d-aeac-449b-bd1f-b8d590ab5ecc-config-data-custom\") pod \"barbican-keystone-listener-6b65859cc8-zkpcj\" (UID: \"5941c09d-aeac-449b-bd1f-b8d590ab5ecc\") " pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.003041 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8976d4b6b-jvhlw"] Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.017943 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-798d46d59c-cpcws"] Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.019958 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.020027 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.020539 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bvnk\" (UniqueName: \"kubernetes.io/projected/2ba68cea-6add-4e52-96ad-458f708a02e0-kube-api-access-4bvnk\") pod \"barbican-worker-5644b469ff-dnc8z\" (UID: \"2ba68cea-6add-4e52-96ad-458f708a02e0\") " pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.036177 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5644b469ff-dnc8z" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.072524 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.105585 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-798d46d59c-cpcws"] Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.141174 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww96x\" (UniqueName: \"kubernetes.io/projected/e9378a8e-48ff-43a6-a1e5-3573f671580b-kube-api-access-ww96x\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.141245 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-combined-ca-bundle\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.141275 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-config-data-custom\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.141298 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44tpl\" (UniqueName: \"kubernetes.io/projected/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-kube-api-access-44tpl\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.141330 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-dns-svc\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.141380 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-config-data\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.141411 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-config\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.141465 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-ovsdbserver-sb\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.141485 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-logs\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.141528 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-ovsdbserver-nb\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.242957 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-config-data\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.243028 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-config\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.243116 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-ovsdbserver-sb\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.243132 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-logs\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.243162 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-ovsdbserver-nb\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.243212 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww96x\" (UniqueName: \"kubernetes.io/projected/e9378a8e-48ff-43a6-a1e5-3573f671580b-kube-api-access-ww96x\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.243275 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-combined-ca-bundle\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.243301 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-config-data-custom\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.243321 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44tpl\" (UniqueName: \"kubernetes.io/projected/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-kube-api-access-44tpl\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.243369 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-dns-svc\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.244510 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-dns-svc\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.245055 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-ovsdbserver-nb\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.245585 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-ovsdbserver-sb\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.245870 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-logs\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.248702 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-config\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.250505 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-config-data-custom\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.250736 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-combined-ca-bundle\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.259730 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-config-data\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.271053 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44tpl\" (UniqueName: \"kubernetes.io/projected/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-kube-api-access-44tpl\") pod \"barbican-api-8976d4b6b-jvhlw\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.295122 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww96x\" (UniqueName: \"kubernetes.io/projected/e9378a8e-48ff-43a6-a1e5-3573f671580b-kube-api-access-ww96x\") pod \"dnsmasq-dns-798d46d59c-cpcws\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.321897 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.371490 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.389685 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dn48v" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.451770 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dn48v" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.451926 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dn48v" event={"ID":"02ce26c3-6755-479e-93c4-295ef8a3ba2a","Type":"ContainerDied","Data":"fafbb22a079ec990b25ece1736036e5e20b70806cd25a05ad9895d314a16af09"} Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.451945 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fafbb22a079ec990b25ece1736036e5e20b70806cd25a05ad9895d314a16af09" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.562776 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-db-sync-config-data\") pod \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.563115 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02ce26c3-6755-479e-93c4-295ef8a3ba2a-etc-machine-id\") pod \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.563188 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02ce26c3-6755-479e-93c4-295ef8a3ba2a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "02ce26c3-6755-479e-93c4-295ef8a3ba2a" (UID: "02ce26c3-6755-479e-93c4-295ef8a3ba2a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.563253 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4667\" (UniqueName: \"kubernetes.io/projected/02ce26c3-6755-479e-93c4-295ef8a3ba2a-kube-api-access-m4667\") pod \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.563281 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-config-data\") pod \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.563351 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-scripts\") pod \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.563368 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-combined-ca-bundle\") pod \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\" (UID: \"02ce26c3-6755-479e-93c4-295ef8a3ba2a\") " Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.567868 4931 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/02ce26c3-6755-479e-93c4-295ef8a3ba2a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.577240 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-scripts" (OuterVolumeSpecName: "scripts") pod "02ce26c3-6755-479e-93c4-295ef8a3ba2a" (UID: "02ce26c3-6755-479e-93c4-295ef8a3ba2a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.577443 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02ce26c3-6755-479e-93c4-295ef8a3ba2a-kube-api-access-m4667" (OuterVolumeSpecName: "kube-api-access-m4667") pod "02ce26c3-6755-479e-93c4-295ef8a3ba2a" (UID: "02ce26c3-6755-479e-93c4-295ef8a3ba2a"). InnerVolumeSpecName "kube-api-access-m4667". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.589135 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "02ce26c3-6755-479e-93c4-295ef8a3ba2a" (UID: "02ce26c3-6755-479e-93c4-295ef8a3ba2a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.629941 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5644b469ff-dnc8z"] Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.634319 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02ce26c3-6755-479e-93c4-295ef8a3ba2a" (UID: "02ce26c3-6755-479e-93c4-295ef8a3ba2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.669699 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4667\" (UniqueName: \"kubernetes.io/projected/02ce26c3-6755-479e-93c4-295ef8a3ba2a-kube-api-access-m4667\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.669730 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.669740 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.669748 4931 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.674536 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-config-data" (OuterVolumeSpecName: "config-data") pod "02ce26c3-6755-479e-93c4-295ef8a3ba2a" (UID: "02ce26c3-6755-479e-93c4-295ef8a3ba2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.767375 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6b65859cc8-zkpcj"] Nov 29 00:19:28 crc kubenswrapper[4931]: I1129 00:19:28.770946 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ce26c3-6755-479e-93c4-295ef8a3ba2a-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:28 crc kubenswrapper[4931]: W1129 00:19:28.788394 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5941c09d_aeac_449b_bd1f_b8d590ab5ecc.slice/crio-820302105e5f9b10b712dc88fe32e7c3fc96c9668e870d3ac089a4c4ddff07f1 WatchSource:0}: Error finding container 820302105e5f9b10b712dc88fe32e7c3fc96c9668e870d3ac089a4c4ddff07f1: Status 404 returned error can't find the container with id 820302105e5f9b10b712dc88fe32e7c3fc96c9668e870d3ac089a4c4ddff07f1 Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.004678 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-798d46d59c-cpcws"] Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.024737 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8976d4b6b-jvhlw"] Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.470563 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" event={"ID":"5941c09d-aeac-449b-bd1f-b8d590ab5ecc","Type":"ContainerStarted","Data":"820302105e5f9b10b712dc88fe32e7c3fc96c9668e870d3ac089a4c4ddff07f1"} Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.473257 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5644b469ff-dnc8z" event={"ID":"2ba68cea-6add-4e52-96ad-458f708a02e0","Type":"ContainerStarted","Data":"88e56c157c59d58b2d61d35f30320ee17573628f7dbc0439cfdaab00a628fddd"} Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.483385 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8976d4b6b-jvhlw" event={"ID":"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8","Type":"ContainerStarted","Data":"972ddf9f7dd4c2d2f4b1631eb8207384e9f00326ad71c1534266ebec8b39a7c9"} Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.483439 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8976d4b6b-jvhlw" event={"ID":"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8","Type":"ContainerStarted","Data":"68d02929f2e8a85d65ec9d44ccb6d2166597a751c2a86307224a00d15a88124d"} Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.487320 4931 generic.go:334] "Generic (PLEG): container finished" podID="e9378a8e-48ff-43a6-a1e5-3573f671580b" containerID="15922055da9132cb0a3aa855f170ba107036fded8a8690b402d86b172110574c" exitCode=0 Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.488668 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798d46d59c-cpcws" event={"ID":"e9378a8e-48ff-43a6-a1e5-3573f671580b","Type":"ContainerDied","Data":"15922055da9132cb0a3aa855f170ba107036fded8a8690b402d86b172110574c"} Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.488697 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798d46d59c-cpcws" event={"ID":"e9378a8e-48ff-43a6-a1e5-3573f671580b","Type":"ContainerStarted","Data":"a4f0f441ddcd047c44204587de9ccebff71fdd9d28125fb5c1337866e41a0adf"} Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.489226 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" podUID="e845944b-9a05-4246-8613-ce413b638e7d" containerName="dnsmasq-dns" containerID="cri-o://b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af" gracePeriod=10 Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.668254 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 00:19:29 crc kubenswrapper[4931]: E1129 00:19:29.674134 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ce26c3-6755-479e-93c4-295ef8a3ba2a" containerName="cinder-db-sync" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.674155 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ce26c3-6755-479e-93c4-295ef8a3ba2a" containerName="cinder-db-sync" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.680511 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="02ce26c3-6755-479e-93c4-295ef8a3ba2a" containerName="cinder-db-sync" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.697209 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.698601 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.705179 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7lgn7" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.705370 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.705501 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.707137 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.758565 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-798d46d59c-cpcws"] Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.799516 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7mj9\" (UniqueName: \"kubernetes.io/projected/b67e3b5f-e942-4719-ae24-bd98ae698859-kube-api-access-q7mj9\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.799599 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b67e3b5f-e942-4719-ae24-bd98ae698859-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.799620 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-config-data\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.799641 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.799685 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.799754 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-scripts\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.824327 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.825892 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.831168 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.859188 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77c9c856fc-g4tr2"] Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.860662 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.897072 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.903185 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b67e3b5f-e942-4719-ae24-bd98ae698859-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.903124 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b67e3b5f-e942-4719-ae24-bd98ae698859-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.903261 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-config-data\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.903288 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.903344 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.903437 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-scripts\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.903507 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7mj9\" (UniqueName: \"kubernetes.io/projected/b67e3b5f-e942-4719-ae24-bd98ae698859-kube-api-access-q7mj9\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.921251 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.922498 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-scripts\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.923279 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-config-data\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.924571 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.933203 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7mj9\" (UniqueName: \"kubernetes.io/projected/b67e3b5f-e942-4719-ae24-bd98ae698859-kube-api-access-q7mj9\") pod \"cinder-scheduler-0\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:29 crc kubenswrapper[4931]: I1129 00:19:29.935894 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77c9c856fc-g4tr2"] Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.004941 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.004992 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-ovsdbserver-nb\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.005018 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-config\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.005053 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-scripts\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.005094 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lndz\" (UniqueName: \"kubernetes.io/projected/7cac2862-eb11-4736-b9cb-68f5c99fdffb-kube-api-access-5lndz\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.005113 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h9mr\" (UniqueName: \"kubernetes.io/projected/f4b84293-3b20-4e73-991b-d3d9fb122446-kube-api-access-6h9mr\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.005159 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-config-data-custom\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.005177 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4b84293-3b20-4e73-991b-d3d9fb122446-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.005201 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-dns-svc\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.005231 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b84293-3b20-4e73-991b-d3d9fb122446-logs\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.005270 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-config-data\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.005290 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-ovsdbserver-sb\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.059289 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111396 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h9mr\" (UniqueName: \"kubernetes.io/projected/f4b84293-3b20-4e73-991b-d3d9fb122446-kube-api-access-6h9mr\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111488 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-config-data-custom\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111508 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4b84293-3b20-4e73-991b-d3d9fb122446-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111536 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-dns-svc\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111563 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b84293-3b20-4e73-991b-d3d9fb122446-logs\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111631 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-config-data\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111648 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-ovsdbserver-sb\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111682 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111701 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-ovsdbserver-nb\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111722 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-config\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111741 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4b84293-3b20-4e73-991b-d3d9fb122446-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111754 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-scripts\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.111902 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lndz\" (UniqueName: \"kubernetes.io/projected/7cac2862-eb11-4736-b9cb-68f5c99fdffb-kube-api-access-5lndz\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.112924 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-ovsdbserver-nb\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.112988 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-ovsdbserver-sb\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.113990 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b84293-3b20-4e73-991b-d3d9fb122446-logs\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.114776 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-config\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.119673 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.120378 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-config-data\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.120442 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-config-data-custom\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.132662 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-dns-svc\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.134075 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-scripts\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.137105 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h9mr\" (UniqueName: \"kubernetes.io/projected/f4b84293-3b20-4e73-991b-d3d9fb122446-kube-api-access-6h9mr\") pod \"cinder-api-0\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.149271 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lndz\" (UniqueName: \"kubernetes.io/projected/7cac2862-eb11-4736-b9cb-68f5c99fdffb-kube-api-access-5lndz\") pod \"dnsmasq-dns-77c9c856fc-g4tr2\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.181873 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.206412 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.397524 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.512822 4931 generic.go:334] "Generic (PLEG): container finished" podID="e845944b-9a05-4246-8613-ce413b638e7d" containerID="b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af" exitCode=0 Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.513144 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" event={"ID":"e845944b-9a05-4246-8613-ce413b638e7d","Type":"ContainerDied","Data":"b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af"} Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.513174 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" event={"ID":"e845944b-9a05-4246-8613-ce413b638e7d","Type":"ContainerDied","Data":"c2caa3e407d3792926852a99a95cf7804a335bfdbfcfb6fb1c434eb8cb45592c"} Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.513192 4931 scope.go:117] "RemoveContainer" containerID="b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.513315 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6c948c7-rjvnp" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.519783 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-ovsdbserver-nb\") pod \"e845944b-9a05-4246-8613-ce413b638e7d\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.520035 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-dns-svc\") pod \"e845944b-9a05-4246-8613-ce413b638e7d\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.520119 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm6dg\" (UniqueName: \"kubernetes.io/projected/e845944b-9a05-4246-8613-ce413b638e7d-kube-api-access-nm6dg\") pod \"e845944b-9a05-4246-8613-ce413b638e7d\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.520270 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-config\") pod \"e845944b-9a05-4246-8613-ce413b638e7d\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.520346 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-ovsdbserver-sb\") pod \"e845944b-9a05-4246-8613-ce413b638e7d\" (UID: \"e845944b-9a05-4246-8613-ce413b638e7d\") " Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.526927 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e845944b-9a05-4246-8613-ce413b638e7d-kube-api-access-nm6dg" (OuterVolumeSpecName: "kube-api-access-nm6dg") pod "e845944b-9a05-4246-8613-ce413b638e7d" (UID: "e845944b-9a05-4246-8613-ce413b638e7d"). InnerVolumeSpecName "kube-api-access-nm6dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.548689 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8976d4b6b-jvhlw" event={"ID":"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8","Type":"ContainerStarted","Data":"58f6af2ae126f9a0e10c3789cabc10d578d65da09809eff0eb725fa63005b3f3"} Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.549931 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.549965 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.582529 4931 scope.go:117] "RemoveContainer" containerID="1068e27541527caaacb8cf808b4362cf4d0d8313d1833865665b72eb246618c7" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.610957 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798d46d59c-cpcws" event={"ID":"e9378a8e-48ff-43a6-a1e5-3573f671580b","Type":"ContainerStarted","Data":"ff409673e49b288522f2ba02dc95cf3b4983afd5ccab7dd7e722d5b7c43f745d"} Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.611098 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-798d46d59c-cpcws" podUID="e9378a8e-48ff-43a6-a1e5-3573f671580b" containerName="dnsmasq-dns" containerID="cri-o://ff409673e49b288522f2ba02dc95cf3b4983afd5ccab7dd7e722d5b7c43f745d" gracePeriod=10 Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.611448 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.622607 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm6dg\" (UniqueName: \"kubernetes.io/projected/e845944b-9a05-4246-8613-ce413b638e7d-kube-api-access-nm6dg\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.635305 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8976d4b6b-jvhlw" podStartSLOduration=3.63528828 podStartE2EDuration="3.63528828s" podCreationTimestamp="2025-11-29 00:19:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:30.598266436 +0000 UTC m=+1183.760159668" watchObservedRunningTime="2025-11-29 00:19:30.63528828 +0000 UTC m=+1183.797181512" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.642948 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-798d46d59c-cpcws" podStartSLOduration=3.64293642 podStartE2EDuration="3.64293642s" podCreationTimestamp="2025-11-29 00:19:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:30.636062653 +0000 UTC m=+1183.797955895" watchObservedRunningTime="2025-11-29 00:19:30.64293642 +0000 UTC m=+1183.804829652" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.647381 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e845944b-9a05-4246-8613-ce413b638e7d" (UID: "e845944b-9a05-4246-8613-ce413b638e7d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.668485 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e845944b-9a05-4246-8613-ce413b638e7d" (UID: "e845944b-9a05-4246-8613-ce413b638e7d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.709063 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e845944b-9a05-4246-8613-ce413b638e7d" (UID: "e845944b-9a05-4246-8613-ce413b638e7d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.709093 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-config" (OuterVolumeSpecName: "config") pod "e845944b-9a05-4246-8613-ce413b638e7d" (UID: "e845944b-9a05-4246-8613-ce413b638e7d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.718998 4931 scope.go:117] "RemoveContainer" containerID="b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af" Nov 29 00:19:30 crc kubenswrapper[4931]: E1129 00:19:30.722933 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af\": container with ID starting with b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af not found: ID does not exist" containerID="b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.722977 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af"} err="failed to get container status \"b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af\": rpc error: code = NotFound desc = could not find container \"b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af\": container with ID starting with b9bc78223a530ba82f1eeae44d8097691593389dbe11c0cb8ddcd0f6776fd6af not found: ID does not exist" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.722998 4931 scope.go:117] "RemoveContainer" containerID="1068e27541527caaacb8cf808b4362cf4d0d8313d1833865665b72eb246618c7" Nov 29 00:19:30 crc kubenswrapper[4931]: E1129 00:19:30.724255 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1068e27541527caaacb8cf808b4362cf4d0d8313d1833865665b72eb246618c7\": container with ID starting with 1068e27541527caaacb8cf808b4362cf4d0d8313d1833865665b72eb246618c7 not found: ID does not exist" containerID="1068e27541527caaacb8cf808b4362cf4d0d8313d1833865665b72eb246618c7" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.724277 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1068e27541527caaacb8cf808b4362cf4d0d8313d1833865665b72eb246618c7"} err="failed to get container status \"1068e27541527caaacb8cf808b4362cf4d0d8313d1833865665b72eb246618c7\": rpc error: code = NotFound desc = could not find container \"1068e27541527caaacb8cf808b4362cf4d0d8313d1833865665b72eb246618c7\": container with ID starting with 1068e27541527caaacb8cf808b4362cf4d0d8313d1833865665b72eb246618c7 not found: ID does not exist" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.725401 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.725429 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.725437 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.725445 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e845944b-9a05-4246-8613-ce413b638e7d-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.861295 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.869726 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6c948c7-rjvnp"] Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.880489 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b6c948c7-rjvnp"] Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.966926 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77c9c856fc-g4tr2"] Nov 29 00:19:30 crc kubenswrapper[4931]: I1129 00:19:30.975617 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 29 00:19:30 crc kubenswrapper[4931]: W1129 00:19:30.981358 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4b84293_3b20_4e73_991b_d3d9fb122446.slice/crio-274f1aa7f5525204964196c5a29d09479c27922a475d0ed0f24d17f9fd3be4ca WatchSource:0}: Error finding container 274f1aa7f5525204964196c5a29d09479c27922a475d0ed0f24d17f9fd3be4ca: Status 404 returned error can't find the container with id 274f1aa7f5525204964196c5a29d09479c27922a475d0ed0f24d17f9fd3be4ca Nov 29 00:19:31 crc kubenswrapper[4931]: I1129 00:19:31.225488 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e845944b-9a05-4246-8613-ce413b638e7d" path="/var/lib/kubelet/pods/e845944b-9a05-4246-8613-ce413b638e7d/volumes" Nov 29 00:19:31 crc kubenswrapper[4931]: I1129 00:19:31.344378 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-765c4d856b-6tkcn" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Nov 29 00:19:31 crc kubenswrapper[4931]: I1129 00:19:31.459505 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-79d99cf788-wwqzd" podUID="4003b615-a779-45a5-b95b-913c80226f4a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Nov 29 00:19:31 crc kubenswrapper[4931]: I1129 00:19:31.660464 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" event={"ID":"7cac2862-eb11-4736-b9cb-68f5c99fdffb","Type":"ContainerStarted","Data":"0434d11badb46d2e34d22777e0051fab4f20819819b17e710b04c954b090d43c"} Nov 29 00:19:31 crc kubenswrapper[4931]: I1129 00:19:31.662127 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b67e3b5f-e942-4719-ae24-bd98ae698859","Type":"ContainerStarted","Data":"e6f9cc02dfc02e085e5d790594bcb6e8cddfbfd587001f05eaa649eeba198b85"} Nov 29 00:19:31 crc kubenswrapper[4931]: I1129 00:19:31.669492 4931 generic.go:334] "Generic (PLEG): container finished" podID="e9378a8e-48ff-43a6-a1e5-3573f671580b" containerID="ff409673e49b288522f2ba02dc95cf3b4983afd5ccab7dd7e722d5b7c43f745d" exitCode=0 Nov 29 00:19:31 crc kubenswrapper[4931]: I1129 00:19:31.669562 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798d46d59c-cpcws" event={"ID":"e9378a8e-48ff-43a6-a1e5-3573f671580b","Type":"ContainerDied","Data":"ff409673e49b288522f2ba02dc95cf3b4983afd5ccab7dd7e722d5b7c43f745d"} Nov 29 00:19:31 crc kubenswrapper[4931]: I1129 00:19:31.671939 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f4b84293-3b20-4e73-991b-d3d9fb122446","Type":"ContainerStarted","Data":"274f1aa7f5525204964196c5a29d09479c27922a475d0ed0f24d17f9fd3be4ca"} Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.296311 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.366151 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-config\") pod \"e9378a8e-48ff-43a6-a1e5-3573f671580b\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.366236 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww96x\" (UniqueName: \"kubernetes.io/projected/e9378a8e-48ff-43a6-a1e5-3573f671580b-kube-api-access-ww96x\") pod \"e9378a8e-48ff-43a6-a1e5-3573f671580b\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.366351 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-ovsdbserver-sb\") pod \"e9378a8e-48ff-43a6-a1e5-3573f671580b\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.366407 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-ovsdbserver-nb\") pod \"e9378a8e-48ff-43a6-a1e5-3573f671580b\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.366433 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-dns-svc\") pod \"e9378a8e-48ff-43a6-a1e5-3573f671580b\" (UID: \"e9378a8e-48ff-43a6-a1e5-3573f671580b\") " Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.380967 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9378a8e-48ff-43a6-a1e5-3573f671580b-kube-api-access-ww96x" (OuterVolumeSpecName: "kube-api-access-ww96x") pod "e9378a8e-48ff-43a6-a1e5-3573f671580b" (UID: "e9378a8e-48ff-43a6-a1e5-3573f671580b"). InnerVolumeSpecName "kube-api-access-ww96x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.425219 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e9378a8e-48ff-43a6-a1e5-3573f671580b" (UID: "e9378a8e-48ff-43a6-a1e5-3573f671580b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.433284 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e9378a8e-48ff-43a6-a1e5-3573f671580b" (UID: "e9378a8e-48ff-43a6-a1e5-3573f671580b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.452775 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-config" (OuterVolumeSpecName: "config") pod "e9378a8e-48ff-43a6-a1e5-3573f671580b" (UID: "e9378a8e-48ff-43a6-a1e5-3573f671580b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.468548 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.468577 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.468588 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.468597 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww96x\" (UniqueName: \"kubernetes.io/projected/e9378a8e-48ff-43a6-a1e5-3573f671580b-kube-api-access-ww96x\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.513424 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e9378a8e-48ff-43a6-a1e5-3573f671580b" (UID: "e9378a8e-48ff-43a6-a1e5-3573f671580b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.570009 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9378a8e-48ff-43a6-a1e5-3573f671580b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.699892 4931 generic.go:334] "Generic (PLEG): container finished" podID="7cac2862-eb11-4736-b9cb-68f5c99fdffb" containerID="25b38f4b06a00bda834703de4f226534cc327513f90975a574b9966dc92c1ae9" exitCode=0 Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.699932 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" event={"ID":"7cac2862-eb11-4736-b9cb-68f5c99fdffb","Type":"ContainerDied","Data":"25b38f4b06a00bda834703de4f226534cc327513f90975a574b9966dc92c1ae9"} Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.703108 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798d46d59c-cpcws" event={"ID":"e9378a8e-48ff-43a6-a1e5-3573f671580b","Type":"ContainerDied","Data":"a4f0f441ddcd047c44204587de9ccebff71fdd9d28125fb5c1337866e41a0adf"} Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.703155 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798d46d59c-cpcws" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.703159 4931 scope.go:117] "RemoveContainer" containerID="ff409673e49b288522f2ba02dc95cf3b4983afd5ccab7dd7e722d5b7c43f745d" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.705958 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f4b84293-3b20-4e73-991b-d3d9fb122446","Type":"ContainerStarted","Data":"4361655e1b14e16992c7f32b972934735282bcf9696745eb490c789afcfe9340"} Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.750723 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-798d46d59c-cpcws"] Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.766050 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-798d46d59c-cpcws"] Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.897730 4931 scope.go:117] "RemoveContainer" containerID="15922055da9132cb0a3aa855f170ba107036fded8a8690b402d86b172110574c" Nov 29 00:19:32 crc kubenswrapper[4931]: I1129 00:19:32.963149 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 29 00:19:33 crc kubenswrapper[4931]: I1129 00:19:33.241803 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9378a8e-48ff-43a6-a1e5-3573f671580b" path="/var/lib/kubelet/pods/e9378a8e-48ff-43a6-a1e5-3573f671580b/volumes" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.492112 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7d5fd9f77d-c4gvt"] Nov 29 00:19:34 crc kubenswrapper[4931]: E1129 00:19:34.493116 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e845944b-9a05-4246-8613-ce413b638e7d" containerName="dnsmasq-dns" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.493132 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e845944b-9a05-4246-8613-ce413b638e7d" containerName="dnsmasq-dns" Nov 29 00:19:34 crc kubenswrapper[4931]: E1129 00:19:34.493155 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9378a8e-48ff-43a6-a1e5-3573f671580b" containerName="init" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.493163 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9378a8e-48ff-43a6-a1e5-3573f671580b" containerName="init" Nov 29 00:19:34 crc kubenswrapper[4931]: E1129 00:19:34.493180 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e845944b-9a05-4246-8613-ce413b638e7d" containerName="init" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.493188 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e845944b-9a05-4246-8613-ce413b638e7d" containerName="init" Nov 29 00:19:34 crc kubenswrapper[4931]: E1129 00:19:34.493222 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9378a8e-48ff-43a6-a1e5-3573f671580b" containerName="dnsmasq-dns" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.493230 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9378a8e-48ff-43a6-a1e5-3573f671580b" containerName="dnsmasq-dns" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.493442 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="e845944b-9a05-4246-8613-ce413b638e7d" containerName="dnsmasq-dns" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.493466 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9378a8e-48ff-43a6-a1e5-3573f671580b" containerName="dnsmasq-dns" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.498768 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.504399 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.504693 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.517063 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7d5fd9f77d-c4gvt"] Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.625397 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-logs\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.625455 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-config-data\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.625493 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq5mn\" (UniqueName: \"kubernetes.io/projected/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-kube-api-access-kq5mn\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.625743 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-config-data-custom\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.625836 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-internal-tls-certs\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.625944 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-combined-ca-bundle\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.626159 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-public-tls-certs\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.727860 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-public-tls-certs\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.727933 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-logs\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.727958 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-config-data\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.727987 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq5mn\" (UniqueName: \"kubernetes.io/projected/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-kube-api-access-kq5mn\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.728071 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-config-data-custom\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.728107 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-internal-tls-certs\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.728157 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-combined-ca-bundle\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.728715 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-logs\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.734177 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-internal-tls-certs\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.738845 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-config-data-custom\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.740065 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-combined-ca-bundle\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.743123 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-config-data\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.756596 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-public-tls-certs\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.760385 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq5mn\" (UniqueName: \"kubernetes.io/projected/d6aa6c84-7c6c-488e-bf3b-7393ad629ed4-kube-api-access-kq5mn\") pod \"barbican-api-7d5fd9f77d-c4gvt\" (UID: \"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4\") " pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.760598 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5644b469ff-dnc8z" event={"ID":"2ba68cea-6add-4e52-96ad-458f708a02e0","Type":"ContainerStarted","Data":"542037340fcd20955dee13c289578c6c7a7725fc50259bcc4482d6a426713ebf"} Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.771184 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" event={"ID":"5941c09d-aeac-449b-bd1f-b8d590ab5ecc","Type":"ContainerStarted","Data":"bbf9a73a77325e7fefcadf087d8ffbf2fbbe2e7f282439991ee8085a4eb7e9e7"} Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.786273 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" event={"ID":"7cac2862-eb11-4736-b9cb-68f5c99fdffb","Type":"ContainerStarted","Data":"1d222546697e9082847b7137f585111ab5a4f8930ca3736f92eb49013870090f"} Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.786319 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.813890 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" podStartSLOduration=5.813869565 podStartE2EDuration="5.813869565s" podCreationTimestamp="2025-11-29 00:19:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:34.811179538 +0000 UTC m=+1187.973072780" watchObservedRunningTime="2025-11-29 00:19:34.813869565 +0000 UTC m=+1187.975762817" Nov 29 00:19:34 crc kubenswrapper[4931]: I1129 00:19:34.824311 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:35 crc kubenswrapper[4931]: I1129 00:19:35.453111 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 29 00:19:35 crc kubenswrapper[4931]: I1129 00:19:35.454691 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 29 00:19:35 crc kubenswrapper[4931]: I1129 00:19:35.502534 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 29 00:19:35 crc kubenswrapper[4931]: I1129 00:19:35.517061 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 29 00:19:35 crc kubenswrapper[4931]: I1129 00:19:35.800625 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b67e3b5f-e942-4719-ae24-bd98ae698859","Type":"ContainerStarted","Data":"fc9ffbd2c9acb232861dc0bdfce3b796cdfed7bada972b6b4068b8812a50ce69"} Nov 29 00:19:35 crc kubenswrapper[4931]: I1129 00:19:35.812552 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f4b84293-3b20-4e73-991b-d3d9fb122446","Type":"ContainerStarted","Data":"6ecd7c6df0749569c7745acd805cf0f8fa14b3e5da166fdf7795875f3a455a4f"} Nov 29 00:19:35 crc kubenswrapper[4931]: I1129 00:19:35.812951 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 29 00:19:35 crc kubenswrapper[4931]: I1129 00:19:35.812971 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 29 00:19:35 crc kubenswrapper[4931]: I1129 00:19:35.813031 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f4b84293-3b20-4e73-991b-d3d9fb122446" containerName="cinder-api-log" containerID="cri-o://4361655e1b14e16992c7f32b972934735282bcf9696745eb490c789afcfe9340" gracePeriod=30 Nov 29 00:19:35 crc kubenswrapper[4931]: I1129 00:19:35.813122 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f4b84293-3b20-4e73-991b-d3d9fb122446" containerName="cinder-api" containerID="cri-o://6ecd7c6df0749569c7745acd805cf0f8fa14b3e5da166fdf7795875f3a455a4f" gracePeriod=30 Nov 29 00:19:35 crc kubenswrapper[4931]: I1129 00:19:35.856564 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.8565427979999996 podStartE2EDuration="6.856542798s" podCreationTimestamp="2025-11-29 00:19:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:35.848788105 +0000 UTC m=+1189.010681357" watchObservedRunningTime="2025-11-29 00:19:35.856542798 +0000 UTC m=+1189.018436030" Nov 29 00:19:36 crc kubenswrapper[4931]: I1129 00:19:36.825857 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b84293-3b20-4e73-991b-d3d9fb122446" containerID="6ecd7c6df0749569c7745acd805cf0f8fa14b3e5da166fdf7795875f3a455a4f" exitCode=0 Nov 29 00:19:36 crc kubenswrapper[4931]: I1129 00:19:36.826979 4931 generic.go:334] "Generic (PLEG): container finished" podID="f4b84293-3b20-4e73-991b-d3d9fb122446" containerID="4361655e1b14e16992c7f32b972934735282bcf9696745eb490c789afcfe9340" exitCode=143 Nov 29 00:19:36 crc kubenswrapper[4931]: I1129 00:19:36.825908 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f4b84293-3b20-4e73-991b-d3d9fb122446","Type":"ContainerDied","Data":"6ecd7c6df0749569c7745acd805cf0f8fa14b3e5da166fdf7795875f3a455a4f"} Nov 29 00:19:36 crc kubenswrapper[4931]: I1129 00:19:36.827575 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f4b84293-3b20-4e73-991b-d3d9fb122446","Type":"ContainerDied","Data":"4361655e1b14e16992c7f32b972934735282bcf9696745eb490c789afcfe9340"} Nov 29 00:19:36 crc kubenswrapper[4931]: I1129 00:19:36.865504 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:36 crc kubenswrapper[4931]: I1129 00:19:36.865550 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:36 crc kubenswrapper[4931]: I1129 00:19:36.912010 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:36 crc kubenswrapper[4931]: I1129 00:19:36.919341 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:37 crc kubenswrapper[4931]: I1129 00:19:37.838674 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:37 crc kubenswrapper[4931]: I1129 00:19:37.839365 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:38 crc kubenswrapper[4931]: I1129 00:19:38.744960 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 29 00:19:38 crc kubenswrapper[4931]: I1129 00:19:38.745277 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 29 00:19:38 crc kubenswrapper[4931]: I1129 00:19:38.884534 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 29 00:19:39 crc kubenswrapper[4931]: I1129 00:19:39.919258 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 29 00:19:39 crc kubenswrapper[4931]: I1129 00:19:39.919495 4931 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.056134 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.182904 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.208764 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.257362 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.258161 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-vjrqg"] Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.258379 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" podUID="8029d5ff-62c3-4088-8122-1e87a2773995" containerName="dnsmasq-dns" containerID="cri-o://c52a3a4e3431c27c0e822288f03d88e131f6e68773372dd961028493a848df85" gracePeriod=10 Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.353258 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.496210 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.927338 4931 generic.go:334] "Generic (PLEG): container finished" podID="8029d5ff-62c3-4088-8122-1e87a2773995" containerID="c52a3a4e3431c27c0e822288f03d88e131f6e68773372dd961028493a848df85" exitCode=0 Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.928104 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" event={"ID":"8029d5ff-62c3-4088-8122-1e87a2773995","Type":"ContainerDied","Data":"c52a3a4e3431c27c0e822288f03d88e131f6e68773372dd961028493a848df85"} Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.952195 4931 generic.go:334] "Generic (PLEG): container finished" podID="39f46125-3c80-4fa1-98d2-88155d268fab" containerID="de9118603db7baa81f5d1a801f22d52cc51caf6ed9abfd7300c882443bfe50e0" exitCode=137 Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.952410 4931 generic.go:334] "Generic (PLEG): container finished" podID="39f46125-3c80-4fa1-98d2-88155d268fab" containerID="1a4a4de6d308cbdb34813d3d4f0f7c6e22ccfeb993fa0be055f679b07a30ac33" exitCode=137 Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.953357 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655c5f955f-jh7hn" event={"ID":"39f46125-3c80-4fa1-98d2-88155d268fab","Type":"ContainerDied","Data":"de9118603db7baa81f5d1a801f22d52cc51caf6ed9abfd7300c882443bfe50e0"} Nov 29 00:19:40 crc kubenswrapper[4931]: I1129 00:19:40.953441 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655c5f955f-jh7hn" event={"ID":"39f46125-3c80-4fa1-98d2-88155d268fab","Type":"ContainerDied","Data":"1a4a4de6d308cbdb34813d3d4f0f7c6e22ccfeb993fa0be055f679b07a30ac33"} Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.822158 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.840736 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-scripts\") pod \"f4b84293-3b20-4e73-991b-d3d9fb122446\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.840782 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-combined-ca-bundle\") pod \"f4b84293-3b20-4e73-991b-d3d9fb122446\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.840958 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h9mr\" (UniqueName: \"kubernetes.io/projected/f4b84293-3b20-4e73-991b-d3d9fb122446-kube-api-access-6h9mr\") pod \"f4b84293-3b20-4e73-991b-d3d9fb122446\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.841022 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-config-data\") pod \"f4b84293-3b20-4e73-991b-d3d9fb122446\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.841076 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4b84293-3b20-4e73-991b-d3d9fb122446-etc-machine-id\") pod \"f4b84293-3b20-4e73-991b-d3d9fb122446\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.841107 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-config-data-custom\") pod \"f4b84293-3b20-4e73-991b-d3d9fb122446\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.841160 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b84293-3b20-4e73-991b-d3d9fb122446-logs\") pod \"f4b84293-3b20-4e73-991b-d3d9fb122446\" (UID: \"f4b84293-3b20-4e73-991b-d3d9fb122446\") " Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.842142 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4b84293-3b20-4e73-991b-d3d9fb122446-logs" (OuterVolumeSpecName: "logs") pod "f4b84293-3b20-4e73-991b-d3d9fb122446" (UID: "f4b84293-3b20-4e73-991b-d3d9fb122446"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.843074 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b84293-3b20-4e73-991b-d3d9fb122446-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f4b84293-3b20-4e73-991b-d3d9fb122446" (UID: "f4b84293-3b20-4e73-991b-d3d9fb122446"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.866521 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-scripts" (OuterVolumeSpecName: "scripts") pod "f4b84293-3b20-4e73-991b-d3d9fb122446" (UID: "f4b84293-3b20-4e73-991b-d3d9fb122446"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.878296 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4b84293-3b20-4e73-991b-d3d9fb122446-kube-api-access-6h9mr" (OuterVolumeSpecName: "kube-api-access-6h9mr") pod "f4b84293-3b20-4e73-991b-d3d9fb122446" (UID: "f4b84293-3b20-4e73-991b-d3d9fb122446"). InnerVolumeSpecName "kube-api-access-6h9mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.938769 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f4b84293-3b20-4e73-991b-d3d9fb122446" (UID: "f4b84293-3b20-4e73-991b-d3d9fb122446"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.947693 4931 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4b84293-3b20-4e73-991b-d3d9fb122446-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.947722 4931 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.947734 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b84293-3b20-4e73-991b-d3d9fb122446-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.947743 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.947753 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h9mr\" (UniqueName: \"kubernetes.io/projected/f4b84293-3b20-4e73-991b-d3d9fb122446-kube-api-access-6h9mr\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.963950 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.964975 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f4b84293-3b20-4e73-991b-d3d9fb122446","Type":"ContainerDied","Data":"274f1aa7f5525204964196c5a29d09479c27922a475d0ed0f24d17f9fd3be4ca"} Nov 29 00:19:41 crc kubenswrapper[4931]: I1129 00:19:41.965010 4931 scope.go:117] "RemoveContainer" containerID="6ecd7c6df0749569c7745acd805cf0f8fa14b3e5da166fdf7795875f3a455a4f" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.071033 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4b84293-3b20-4e73-991b-d3d9fb122446" (UID: "f4b84293-3b20-4e73-991b-d3d9fb122446"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.118879 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-config-data" (OuterVolumeSpecName: "config-data") pod "f4b84293-3b20-4e73-991b-d3d9fb122446" (UID: "f4b84293-3b20-4e73-991b-d3d9fb122446"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.164489 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.164526 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b84293-3b20-4e73-991b-d3d9fb122446-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.453178 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.458032 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7d5fd9f77d-c4gvt"] Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.486803 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-ovsdbserver-nb\") pod \"8029d5ff-62c3-4088-8122-1e87a2773995\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.486918 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grt4p\" (UniqueName: \"kubernetes.io/projected/8029d5ff-62c3-4088-8122-1e87a2773995-kube-api-access-grt4p\") pod \"8029d5ff-62c3-4088-8122-1e87a2773995\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.487062 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-dns-svc\") pod \"8029d5ff-62c3-4088-8122-1e87a2773995\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.487150 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-ovsdbserver-sb\") pod \"8029d5ff-62c3-4088-8122-1e87a2773995\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.487179 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-config\") pod \"8029d5ff-62c3-4088-8122-1e87a2773995\" (UID: \"8029d5ff-62c3-4088-8122-1e87a2773995\") " Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.502999 4931 scope.go:117] "RemoveContainer" containerID="4361655e1b14e16992c7f32b972934735282bcf9696745eb490c789afcfe9340" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.521625 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8029d5ff-62c3-4088-8122-1e87a2773995-kube-api-access-grt4p" (OuterVolumeSpecName: "kube-api-access-grt4p") pod "8029d5ff-62c3-4088-8122-1e87a2773995" (UID: "8029d5ff-62c3-4088-8122-1e87a2773995"). InnerVolumeSpecName "kube-api-access-grt4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.526531 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.545256 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.588875 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxbh9\" (UniqueName: \"kubernetes.io/projected/39f46125-3c80-4fa1-98d2-88155d268fab-kube-api-access-xxbh9\") pod \"39f46125-3c80-4fa1-98d2-88155d268fab\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.588976 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39f46125-3c80-4fa1-98d2-88155d268fab-logs\") pod \"39f46125-3c80-4fa1-98d2-88155d268fab\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.589031 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39f46125-3c80-4fa1-98d2-88155d268fab-config-data\") pod \"39f46125-3c80-4fa1-98d2-88155d268fab\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.589048 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39f46125-3c80-4fa1-98d2-88155d268fab-scripts\") pod \"39f46125-3c80-4fa1-98d2-88155d268fab\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.589078 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/39f46125-3c80-4fa1-98d2-88155d268fab-horizon-secret-key\") pod \"39f46125-3c80-4fa1-98d2-88155d268fab\" (UID: \"39f46125-3c80-4fa1-98d2-88155d268fab\") " Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.589491 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grt4p\" (UniqueName: \"kubernetes.io/projected/8029d5ff-62c3-4088-8122-1e87a2773995-kube-api-access-grt4p\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.590011 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39f46125-3c80-4fa1-98d2-88155d268fab-logs" (OuterVolumeSpecName: "logs") pod "39f46125-3c80-4fa1-98d2-88155d268fab" (UID: "39f46125-3c80-4fa1-98d2-88155d268fab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.592155 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.602000 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 29 00:19:42 crc kubenswrapper[4931]: E1129 00:19:42.602399 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39f46125-3c80-4fa1-98d2-88155d268fab" containerName="horizon" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.602416 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="39f46125-3c80-4fa1-98d2-88155d268fab" containerName="horizon" Nov 29 00:19:42 crc kubenswrapper[4931]: E1129 00:19:42.602432 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8029d5ff-62c3-4088-8122-1e87a2773995" containerName="dnsmasq-dns" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.602440 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="8029d5ff-62c3-4088-8122-1e87a2773995" containerName="dnsmasq-dns" Nov 29 00:19:42 crc kubenswrapper[4931]: E1129 00:19:42.602458 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39f46125-3c80-4fa1-98d2-88155d268fab" containerName="horizon-log" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.602464 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="39f46125-3c80-4fa1-98d2-88155d268fab" containerName="horizon-log" Nov 29 00:19:42 crc kubenswrapper[4931]: E1129 00:19:42.602487 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b84293-3b20-4e73-991b-d3d9fb122446" containerName="cinder-api" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.602493 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b84293-3b20-4e73-991b-d3d9fb122446" containerName="cinder-api" Nov 29 00:19:42 crc kubenswrapper[4931]: E1129 00:19:42.602507 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b84293-3b20-4e73-991b-d3d9fb122446" containerName="cinder-api-log" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.602514 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b84293-3b20-4e73-991b-d3d9fb122446" containerName="cinder-api-log" Nov 29 00:19:42 crc kubenswrapper[4931]: E1129 00:19:42.602523 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8029d5ff-62c3-4088-8122-1e87a2773995" containerName="init" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.602529 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="8029d5ff-62c3-4088-8122-1e87a2773995" containerName="init" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.602708 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="39f46125-3c80-4fa1-98d2-88155d268fab" containerName="horizon-log" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.602723 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b84293-3b20-4e73-991b-d3d9fb122446" containerName="cinder-api-log" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.602733 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="8029d5ff-62c3-4088-8122-1e87a2773995" containerName="dnsmasq-dns" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.602744 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b84293-3b20-4e73-991b-d3d9fb122446" containerName="cinder-api" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.602759 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="39f46125-3c80-4fa1-98d2-88155d268fab" containerName="horizon" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.603703 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.609477 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.609992 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.614381 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.615402 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.624292 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39f46125-3c80-4fa1-98d2-88155d268fab-kube-api-access-xxbh9" (OuterVolumeSpecName: "kube-api-access-xxbh9") pod "39f46125-3c80-4fa1-98d2-88155d268fab" (UID: "39f46125-3c80-4fa1-98d2-88155d268fab"). InnerVolumeSpecName "kube-api-access-xxbh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.625088 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39f46125-3c80-4fa1-98d2-88155d268fab-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "39f46125-3c80-4fa1-98d2-88155d268fab" (UID: "39f46125-3c80-4fa1-98d2-88155d268fab"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.691472 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.691847 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.691870 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqvnw\" (UniqueName: \"kubernetes.io/projected/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-kube-api-access-mqvnw\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.691946 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.691976 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-config-data-custom\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.692010 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-logs\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.692038 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-config-data\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.692066 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-scripts\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.692085 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.692131 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxbh9\" (UniqueName: \"kubernetes.io/projected/39f46125-3c80-4fa1-98d2-88155d268fab-kube-api-access-xxbh9\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.692142 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39f46125-3c80-4fa1-98d2-88155d268fab-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.692151 4931 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/39f46125-3c80-4fa1-98d2-88155d268fab-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.746637 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-config" (OuterVolumeSpecName: "config") pod "8029d5ff-62c3-4088-8122-1e87a2773995" (UID: "8029d5ff-62c3-4088-8122-1e87a2773995"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.753359 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39f46125-3c80-4fa1-98d2-88155d268fab-config-data" (OuterVolumeSpecName: "config-data") pod "39f46125-3c80-4fa1-98d2-88155d268fab" (UID: "39f46125-3c80-4fa1-98d2-88155d268fab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.759414 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8029d5ff-62c3-4088-8122-1e87a2773995" (UID: "8029d5ff-62c3-4088-8122-1e87a2773995"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.767474 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8029d5ff-62c3-4088-8122-1e87a2773995" (UID: "8029d5ff-62c3-4088-8122-1e87a2773995"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.779420 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8029d5ff-62c3-4088-8122-1e87a2773995" (UID: "8029d5ff-62c3-4088-8122-1e87a2773995"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.783879 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39f46125-3c80-4fa1-98d2-88155d268fab-scripts" (OuterVolumeSpecName: "scripts") pod "39f46125-3c80-4fa1-98d2-88155d268fab" (UID: "39f46125-3c80-4fa1-98d2-88155d268fab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793593 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793633 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-config-data-custom\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793670 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-logs\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793696 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-config-data\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793726 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-scripts\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793746 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793766 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793820 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793843 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqvnw\" (UniqueName: \"kubernetes.io/projected/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-kube-api-access-mqvnw\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793932 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39f46125-3c80-4fa1-98d2-88155d268fab-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793942 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39f46125-3c80-4fa1-98d2-88155d268fab-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793950 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793960 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793968 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.793976 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8029d5ff-62c3-4088-8122-1e87a2773995-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.796227 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-logs\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.796518 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.798734 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-config-data-custom\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.799096 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-public-tls-certs\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.801885 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.802888 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-config-data\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.803482 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.803893 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-scripts\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.809176 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqvnw\" (UniqueName: \"kubernetes.io/projected/80c8c44d-d0c0-4d7a-9063-9915b1c4aea6-kube-api-access-mqvnw\") pod \"cinder-api-0\" (UID: \"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6\") " pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.931075 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.975791 4931 generic.go:334] "Generic (PLEG): container finished" podID="8826b2d2-afb8-449c-8b42-c3d8e50cfe59" containerID="a58ec12db7bf12bdee74fad4da782c91dc8a2f6eea9b6cf6462f9c481932d58a" exitCode=137 Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.975844 4931 generic.go:334] "Generic (PLEG): container finished" podID="8826b2d2-afb8-449c-8b42-c3d8e50cfe59" containerID="0683ba9e2568040443e68a523b3b1a085df5239d2cd7c5bd50471c51d6091d8a" exitCode=137 Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.975890 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b697487c7-wzjcs" event={"ID":"8826b2d2-afb8-449c-8b42-c3d8e50cfe59","Type":"ContainerDied","Data":"a58ec12db7bf12bdee74fad4da782c91dc8a2f6eea9b6cf6462f9c481932d58a"} Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.975924 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b697487c7-wzjcs" event={"ID":"8826b2d2-afb8-449c-8b42-c3d8e50cfe59","Type":"ContainerDied","Data":"0683ba9e2568040443e68a523b3b1a085df5239d2cd7c5bd50471c51d6091d8a"} Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.977854 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655c5f955f-jh7hn" event={"ID":"39f46125-3c80-4fa1-98d2-88155d268fab","Type":"ContainerDied","Data":"3ca636d89c167be0e9b5a09f3a2781c7ab78c3b260c42cdaf42ef57ee474311c"} Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.977895 4931 scope.go:117] "RemoveContainer" containerID="de9118603db7baa81f5d1a801f22d52cc51caf6ed9abfd7300c882443bfe50e0" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.978016 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655c5f955f-jh7hn" Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.988556 4931 generic.go:334] "Generic (PLEG): container finished" podID="653aa9fc-902d-4186-812f-168b8a822c89" containerID="814dd27c23460a30a37856085da4a4026f6f6a5c5259467c0ea4b7ba5818f758" exitCode=137 Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.988606 4931 generic.go:334] "Generic (PLEG): container finished" podID="653aa9fc-902d-4186-812f-168b8a822c89" containerID="d78e686bf2fd86b74ce23a091bdf60c608d88f5d204a9866f5c892ce4b67221b" exitCode=137 Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.988666 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-784c4db8c5-j7sst" event={"ID":"653aa9fc-902d-4186-812f-168b8a822c89","Type":"ContainerDied","Data":"814dd27c23460a30a37856085da4a4026f6f6a5c5259467c0ea4b7ba5818f758"} Nov 29 00:19:42 crc kubenswrapper[4931]: I1129 00:19:42.988700 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-784c4db8c5-j7sst" event={"ID":"653aa9fc-902d-4186-812f-168b8a822c89","Type":"ContainerDied","Data":"d78e686bf2fd86b74ce23a091bdf60c608d88f5d204a9866f5c892ce4b67221b"} Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.002819 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" event={"ID":"8029d5ff-62c3-4088-8122-1e87a2773995","Type":"ContainerDied","Data":"9d80cb1588ece603329894756695d8faa7931d6bf8e20fb381a0e17edf399e94"} Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.002953 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-vjrqg" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.010696 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d5fd9f77d-c4gvt" event={"ID":"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4","Type":"ContainerStarted","Data":"8750011eeb121aeb7d495ee6bf86580037d7550e8fe32904de645a4fd80abb76"} Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.162865 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-vjrqg"] Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.179882 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-vjrqg"] Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.184857 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-655c5f955f-jh7hn"] Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.190272 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-655c5f955f-jh7hn"] Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.222851 4931 scope.go:117] "RemoveContainer" containerID="1a4a4de6d308cbdb34813d3d4f0f7c6e22ccfeb993fa0be055f679b07a30ac33" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.234076 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39f46125-3c80-4fa1-98d2-88155d268fab" path="/var/lib/kubelet/pods/39f46125-3c80-4fa1-98d2-88155d268fab/volumes" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.234627 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8029d5ff-62c3-4088-8122-1e87a2773995" path="/var/lib/kubelet/pods/8029d5ff-62c3-4088-8122-1e87a2773995/volumes" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.235211 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b84293-3b20-4e73-991b-d3d9fb122446" path="/var/lib/kubelet/pods/f4b84293-3b20-4e73-991b-d3d9fb122446/volumes" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.268565 4931 scope.go:117] "RemoveContainer" containerID="c52a3a4e3431c27c0e822288f03d88e131f6e68773372dd961028493a848df85" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.308414 4931 scope.go:117] "RemoveContainer" containerID="a0b67bbd7e690ad64f6bcb71b626d3bde02533eea0236cf86f4f8ed8e1a57775" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.509336 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.633614 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-config-data\") pod \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.633667 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-horizon-secret-key\") pod \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.633827 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-scripts\") pod \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.633892 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llfvm\" (UniqueName: \"kubernetes.io/projected/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-kube-api-access-llfvm\") pod \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.634658 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-logs\") pod \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\" (UID: \"8826b2d2-afb8-449c-8b42-c3d8e50cfe59\") " Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.635407 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-logs" (OuterVolumeSpecName: "logs") pod "8826b2d2-afb8-449c-8b42-c3d8e50cfe59" (UID: "8826b2d2-afb8-449c-8b42-c3d8e50cfe59"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.640215 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8826b2d2-afb8-449c-8b42-c3d8e50cfe59" (UID: "8826b2d2-afb8-449c-8b42-c3d8e50cfe59"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.649055 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-kube-api-access-llfvm" (OuterVolumeSpecName: "kube-api-access-llfvm") pod "8826b2d2-afb8-449c-8b42-c3d8e50cfe59" (UID: "8826b2d2-afb8-449c-8b42-c3d8e50cfe59"). InnerVolumeSpecName "kube-api-access-llfvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.662422 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.669025 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-scripts" (OuterVolumeSpecName: "scripts") pod "8826b2d2-afb8-449c-8b42-c3d8e50cfe59" (UID: "8826b2d2-afb8-449c-8b42-c3d8e50cfe59"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.698253 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-config-data" (OuterVolumeSpecName: "config-data") pod "8826b2d2-afb8-449c-8b42-c3d8e50cfe59" (UID: "8826b2d2-afb8-449c-8b42-c3d8e50cfe59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.731996 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 29 00:19:43 crc kubenswrapper[4931]: W1129 00:19:43.735307 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80c8c44d_d0c0_4d7a_9063_9915b1c4aea6.slice/crio-e47b903c447d619f838fb0c790476dfac9f00f2e9427069018299579499e3231 WatchSource:0}: Error finding container e47b903c447d619f838fb0c790476dfac9f00f2e9427069018299579499e3231: Status 404 returned error can't find the container with id e47b903c447d619f838fb0c790476dfac9f00f2e9427069018299579499e3231 Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.735964 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/653aa9fc-902d-4186-812f-168b8a822c89-horizon-secret-key\") pod \"653aa9fc-902d-4186-812f-168b8a822c89\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.736007 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwtwm\" (UniqueName: \"kubernetes.io/projected/653aa9fc-902d-4186-812f-168b8a822c89-kube-api-access-dwtwm\") pod \"653aa9fc-902d-4186-812f-168b8a822c89\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.736037 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/653aa9fc-902d-4186-812f-168b8a822c89-config-data\") pod \"653aa9fc-902d-4186-812f-168b8a822c89\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.736156 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/653aa9fc-902d-4186-812f-168b8a822c89-logs\") pod \"653aa9fc-902d-4186-812f-168b8a822c89\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.736339 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/653aa9fc-902d-4186-812f-168b8a822c89-scripts\") pod \"653aa9fc-902d-4186-812f-168b8a822c89\" (UID: \"653aa9fc-902d-4186-812f-168b8a822c89\") " Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.736760 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.736776 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.736786 4931 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.736796 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.736824 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llfvm\" (UniqueName: \"kubernetes.io/projected/8826b2d2-afb8-449c-8b42-c3d8e50cfe59-kube-api-access-llfvm\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.737208 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/653aa9fc-902d-4186-812f-168b8a822c89-logs" (OuterVolumeSpecName: "logs") pod "653aa9fc-902d-4186-812f-168b8a822c89" (UID: "653aa9fc-902d-4186-812f-168b8a822c89"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.743427 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/653aa9fc-902d-4186-812f-168b8a822c89-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "653aa9fc-902d-4186-812f-168b8a822c89" (UID: "653aa9fc-902d-4186-812f-168b8a822c89"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.744395 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/653aa9fc-902d-4186-812f-168b8a822c89-kube-api-access-dwtwm" (OuterVolumeSpecName: "kube-api-access-dwtwm") pod "653aa9fc-902d-4186-812f-168b8a822c89" (UID: "653aa9fc-902d-4186-812f-168b8a822c89"). InnerVolumeSpecName "kube-api-access-dwtwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.779471 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/653aa9fc-902d-4186-812f-168b8a822c89-config-data" (OuterVolumeSpecName: "config-data") pod "653aa9fc-902d-4186-812f-168b8a822c89" (UID: "653aa9fc-902d-4186-812f-168b8a822c89"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.785255 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/653aa9fc-902d-4186-812f-168b8a822c89-scripts" (OuterVolumeSpecName: "scripts") pod "653aa9fc-902d-4186-812f-168b8a822c89" (UID: "653aa9fc-902d-4186-812f-168b8a822c89"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.838050 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/653aa9fc-902d-4186-812f-168b8a822c89-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.838080 4931 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/653aa9fc-902d-4186-812f-168b8a822c89-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.838090 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwtwm\" (UniqueName: \"kubernetes.io/projected/653aa9fc-902d-4186-812f-168b8a822c89-kube-api-access-dwtwm\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.838100 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/653aa9fc-902d-4186-812f-168b8a822c89-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.838108 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/653aa9fc-902d-4186-812f-168b8a822c89-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.890618 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:19:43 crc kubenswrapper[4931]: I1129 00:19:43.967245 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.053161 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b67e3b5f-e942-4719-ae24-bd98ae698859","Type":"ContainerStarted","Data":"76dfdd0136931cb4a46398d2ae3d1e90ecc3611e19ee01719564dcc2523fb0b9"} Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.055864 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d5fd9f77d-c4gvt" event={"ID":"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4","Type":"ContainerStarted","Data":"b848932afc096a2a7d59c23e86b4386e99afe81bb1724a445f07fd4d16f024a6"} Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.059111 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b697487c7-wzjcs" event={"ID":"8826b2d2-afb8-449c-8b42-c3d8e50cfe59","Type":"ContainerDied","Data":"9d18cfef807639acf2ba2c074599af59d96a4b5b216ada3ad40ff7aa4e4d01c6"} Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.059151 4931 scope.go:117] "RemoveContainer" containerID="a58ec12db7bf12bdee74fad4da782c91dc8a2f6eea9b6cf6462f9c481932d58a" Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.059234 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b697487c7-wzjcs" Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.061198 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6","Type":"ContainerStarted","Data":"e47b903c447d619f838fb0c790476dfac9f00f2e9427069018299579499e3231"} Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.063822 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-784c4db8c5-j7sst" event={"ID":"653aa9fc-902d-4186-812f-168b8a822c89","Type":"ContainerDied","Data":"8a412fa3d7d9f02049897ff8a78284ffa73490847c10c37d5f52daa98443db75"} Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.063885 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-784c4db8c5-j7sst" Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.080078 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5644b469ff-dnc8z" event={"ID":"2ba68cea-6add-4e52-96ad-458f708a02e0","Type":"ContainerStarted","Data":"127c2dbbcb1265ef2c1c1b1dc107337c501d81ffb630e8ceb0c13f6f46dda9b4"} Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.083110 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=11.948964562 podStartE2EDuration="15.083094655s" podCreationTimestamp="2025-11-29 00:19:29 +0000 UTC" firstStartedPulling="2025-11-29 00:19:30.883584047 +0000 UTC m=+1184.045477279" lastFinishedPulling="2025-11-29 00:19:34.01771414 +0000 UTC m=+1187.179607372" observedRunningTime="2025-11-29 00:19:44.081792677 +0000 UTC m=+1197.243685919" watchObservedRunningTime="2025-11-29 00:19:44.083094655 +0000 UTC m=+1197.244987887" Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.112349 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57b51201-4e02-4b29-843d-6bfe84247cd8","Type":"ContainerStarted","Data":"39d695aa87a1a566d711b861dccd1c5cd384c808f103dbe32f0c6665e379f6e7"} Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.112622 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="ceilometer-central-agent" containerID="cri-o://bdedfb8b573d44dcd2880bd5f7719721cac9337ba0fe28318bdbf0e5b580ad90" gracePeriod=30 Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.112841 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.112909 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="sg-core" containerID="cri-o://9d01970592fdce0484783d5f757c978cbd0536d5468f72ad79ddae208fd84d3d" gracePeriod=30 Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.113078 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="ceilometer-notification-agent" containerID="cri-o://8e97f61f77404e3bbfb6af14a2ed78fa4f9311b5b8a206712ad11e9b1b2cb0b4" gracePeriod=30 Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.113140 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="proxy-httpd" containerID="cri-o://39d695aa87a1a566d711b861dccd1c5cd384c808f103dbe32f0c6665e379f6e7" gracePeriod=30 Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.132824 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" event={"ID":"5941c09d-aeac-449b-bd1f-b8d590ab5ecc","Type":"ContainerStarted","Data":"53c69da06c82820db60a1cf561aaf73445498bdb6a5dbea0221c4dfd444d6b2d"} Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.175337 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5644b469ff-dnc8z" podStartSLOduration=11.972368434 podStartE2EDuration="17.175316386s" podCreationTimestamp="2025-11-29 00:19:27 +0000 UTC" firstStartedPulling="2025-11-29 00:19:28.680092617 +0000 UTC m=+1181.841985849" lastFinishedPulling="2025-11-29 00:19:33.883040559 +0000 UTC m=+1187.044933801" observedRunningTime="2025-11-29 00:19:44.126085061 +0000 UTC m=+1197.287978293" watchObservedRunningTime="2025-11-29 00:19:44.175316386 +0000 UTC m=+1197.337209618" Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.190031 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-784c4db8c5-j7sst"] Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.201213 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-784c4db8c5-j7sst"] Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.213826 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-b697487c7-wzjcs"] Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.238277 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-b697487c7-wzjcs"] Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.256283 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.190623887 podStartE2EDuration="1m12.256145519s" podCreationTimestamp="2025-11-29 00:18:32 +0000 UTC" firstStartedPulling="2025-11-29 00:18:34.21444511 +0000 UTC m=+1127.376338342" lastFinishedPulling="2025-11-29 00:19:42.279966752 +0000 UTC m=+1195.441859974" observedRunningTime="2025-11-29 00:19:44.203471355 +0000 UTC m=+1197.365364617" watchObservedRunningTime="2025-11-29 00:19:44.256145519 +0000 UTC m=+1197.418038751" Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.277466 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6b65859cc8-zkpcj" podStartSLOduration=12.183189604 podStartE2EDuration="17.277450002s" podCreationTimestamp="2025-11-29 00:19:27 +0000 UTC" firstStartedPulling="2025-11-29 00:19:28.790047347 +0000 UTC m=+1181.951940579" lastFinishedPulling="2025-11-29 00:19:33.884307745 +0000 UTC m=+1187.046200977" observedRunningTime="2025-11-29 00:19:44.22172027 +0000 UTC m=+1197.383613522" watchObservedRunningTime="2025-11-29 00:19:44.277450002 +0000 UTC m=+1197.439343234" Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.351095 4931 scope.go:117] "RemoveContainer" containerID="0683ba9e2568040443e68a523b3b1a085df5239d2cd7c5bd50471c51d6091d8a" Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.529242 4931 scope.go:117] "RemoveContainer" containerID="814dd27c23460a30a37856085da4a4026f6f6a5c5259467c0ea4b7ba5818f758" Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.770752 4931 scope.go:117] "RemoveContainer" containerID="d78e686bf2fd86b74ce23a091bdf60c608d88f5d204a9866f5c892ce4b67221b" Nov 29 00:19:44 crc kubenswrapper[4931]: I1129 00:19:44.872965 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.065993 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.072769 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="b67e3b5f-e942-4719-ae24-bd98ae698859" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.159:8080/\": dial tcp 10.217.0.159:8080: connect: connection refused" Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.169176 4931 generic.go:334] "Generic (PLEG): container finished" podID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerID="39d695aa87a1a566d711b861dccd1c5cd384c808f103dbe32f0c6665e379f6e7" exitCode=0 Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.169212 4931 generic.go:334] "Generic (PLEG): container finished" podID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerID="9d01970592fdce0484783d5f757c978cbd0536d5468f72ad79ddae208fd84d3d" exitCode=2 Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.169220 4931 generic.go:334] "Generic (PLEG): container finished" podID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerID="bdedfb8b573d44dcd2880bd5f7719721cac9337ba0fe28318bdbf0e5b580ad90" exitCode=0 Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.169260 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57b51201-4e02-4b29-843d-6bfe84247cd8","Type":"ContainerDied","Data":"39d695aa87a1a566d711b861dccd1c5cd384c808f103dbe32f0c6665e379f6e7"} Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.169285 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57b51201-4e02-4b29-843d-6bfe84247cd8","Type":"ContainerDied","Data":"9d01970592fdce0484783d5f757c978cbd0536d5468f72ad79ddae208fd84d3d"} Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.169295 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57b51201-4e02-4b29-843d-6bfe84247cd8","Type":"ContainerDied","Data":"bdedfb8b573d44dcd2880bd5f7719721cac9337ba0fe28318bdbf0e5b580ad90"} Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.172776 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6","Type":"ContainerStarted","Data":"782de7dece3771a87ecafc64462bffd5abd83406d68e1965bf557dd5a2bc89df"} Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.179417 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d5fd9f77d-c4gvt" event={"ID":"d6aa6c84-7c6c-488e-bf3b-7393ad629ed4","Type":"ContainerStarted","Data":"452678b90c65cfb7e8a316fa22bb8cbf9e9029bcd82e248a74412d1fe3b545cd"} Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.180927 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.180968 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.183013 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="f4b84293-3b20-4e73-991b-d3d9fb122446" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.160:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.229325 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7d5fd9f77d-c4gvt" podStartSLOduration=11.229309353 podStartE2EDuration="11.229309353s" podCreationTimestamp="2025-11-29 00:19:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:45.222195458 +0000 UTC m=+1198.384088700" watchObservedRunningTime="2025-11-29 00:19:45.229309353 +0000 UTC m=+1198.391202585" Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.234540 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="653aa9fc-902d-4186-812f-168b8a822c89" path="/var/lib/kubelet/pods/653aa9fc-902d-4186-812f-168b8a822c89/volumes" Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.235249 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8826b2d2-afb8-449c-8b42-c3d8e50cfe59" path="/var/lib/kubelet/pods/8826b2d2-afb8-449c-8b42-c3d8e50cfe59/volumes" Nov 29 00:19:45 crc kubenswrapper[4931]: I1129 00:19:45.978214 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:19:46 crc kubenswrapper[4931]: I1129 00:19:46.030524 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-79d99cf788-wwqzd" Nov 29 00:19:46 crc kubenswrapper[4931]: I1129 00:19:46.099909 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-765c4d856b-6tkcn"] Nov 29 00:19:46 crc kubenswrapper[4931]: I1129 00:19:46.193022 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"80c8c44d-d0c0-4d7a-9063-9915b1c4aea6","Type":"ContainerStarted","Data":"770e683d24881d93e7866ea40dae35e1a587d73c235bcd2a18089118e048025d"} Nov 29 00:19:46 crc kubenswrapper[4931]: I1129 00:19:46.194250 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-765c4d856b-6tkcn" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon-log" containerID="cri-o://da9b24f68267e8a0f02647614f89605bc826dc9c40f4d64d967492a86d47c0eb" gracePeriod=30 Nov 29 00:19:46 crc kubenswrapper[4931]: I1129 00:19:46.194513 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 29 00:19:46 crc kubenswrapper[4931]: I1129 00:19:46.194591 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-765c4d856b-6tkcn" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon" containerID="cri-o://44ea44f6e434cd440fd70136642061dd183226c05674df61c397df90f3874531" gracePeriod=30 Nov 29 00:19:47 crc kubenswrapper[4931]: E1129 00:19:47.046742 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-storage-0" podUID="a720d879-8b09-452d-8371-9bf75f28cbde" Nov 29 00:19:47 crc kubenswrapper[4931]: I1129 00:19:47.203653 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 29 00:19:47 crc kubenswrapper[4931]: I1129 00:19:47.915930 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-bcd7d89b9-l4sjs" Nov 29 00:19:47 crc kubenswrapper[4931]: I1129 00:19:47.945298 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.945272286 podStartE2EDuration="5.945272286s" podCreationTimestamp="2025-11-29 00:19:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:19:46.237328469 +0000 UTC m=+1199.399221701" watchObservedRunningTime="2025-11-29 00:19:47.945272286 +0000 UTC m=+1201.107165548" Nov 29 00:19:47 crc kubenswrapper[4931]: I1129 00:19:47.981454 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b8d6898dd-wpzld"] Nov 29 00:19:47 crc kubenswrapper[4931]: I1129 00:19:47.981702 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-b8d6898dd-wpzld" podUID="f7044c78-4dc6-4934-8d91-e8735d23be54" containerName="neutron-api" containerID="cri-o://4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3" gracePeriod=30 Nov 29 00:19:47 crc kubenswrapper[4931]: I1129 00:19:47.982063 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-b8d6898dd-wpzld" podUID="f7044c78-4dc6-4934-8d91-e8735d23be54" containerName="neutron-httpd" containerID="cri-o://7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122" gracePeriod=30 Nov 29 00:19:50 crc kubenswrapper[4931]: I1129 00:19:50.236373 4931 generic.go:334] "Generic (PLEG): container finished" podID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerID="44ea44f6e434cd440fd70136642061dd183226c05674df61c397df90f3874531" exitCode=0 Nov 29 00:19:50 crc kubenswrapper[4931]: I1129 00:19:50.236947 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-765c4d856b-6tkcn" event={"ID":"77f47a1f-3302-4d58-9b94-4ea756cc38ab","Type":"ContainerDied","Data":"44ea44f6e434cd440fd70136642061dd183226c05674df61c397df90f3874531"} Nov 29 00:19:50 crc kubenswrapper[4931]: I1129 00:19:50.264378 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 29 00:19:50 crc kubenswrapper[4931]: I1129 00:19:50.320270 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.255895 4931 generic.go:334] "Generic (PLEG): container finished" podID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerID="8e97f61f77404e3bbfb6af14a2ed78fa4f9311b5b8a206712ad11e9b1b2cb0b4" exitCode=0 Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.256245 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57b51201-4e02-4b29-843d-6bfe84247cd8","Type":"ContainerDied","Data":"8e97f61f77404e3bbfb6af14a2ed78fa4f9311b5b8a206712ad11e9b1b2cb0b4"} Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.258922 4931 generic.go:334] "Generic (PLEG): container finished" podID="f7044c78-4dc6-4934-8d91-e8735d23be54" containerID="7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122" exitCode=0 Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.259003 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b8d6898dd-wpzld" event={"ID":"f7044c78-4dc6-4934-8d91-e8735d23be54","Type":"ContainerDied","Data":"7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122"} Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.259293 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b67e3b5f-e942-4719-ae24-bd98ae698859" containerName="probe" containerID="cri-o://76dfdd0136931cb4a46398d2ae3d1e90ecc3611e19ee01719564dcc2523fb0b9" gracePeriod=30 Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.259451 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b67e3b5f-e942-4719-ae24-bd98ae698859" containerName="cinder-scheduler" containerID="cri-o://fc9ffbd2c9acb232861dc0bdfce3b796cdfed7bada972b6b4068b8812a50ce69" gracePeriod=30 Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.337958 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.344721 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-765c4d856b-6tkcn" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.424146 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b51201-4e02-4b29-843d-6bfe84247cd8-run-httpd\") pod \"57b51201-4e02-4b29-843d-6bfe84247cd8\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.424215 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-scripts\") pod \"57b51201-4e02-4b29-843d-6bfe84247cd8\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.424246 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctvl6\" (UniqueName: \"kubernetes.io/projected/57b51201-4e02-4b29-843d-6bfe84247cd8-kube-api-access-ctvl6\") pod \"57b51201-4e02-4b29-843d-6bfe84247cd8\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.424274 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-combined-ca-bundle\") pod \"57b51201-4e02-4b29-843d-6bfe84247cd8\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.424295 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b51201-4e02-4b29-843d-6bfe84247cd8-log-httpd\") pod \"57b51201-4e02-4b29-843d-6bfe84247cd8\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.424325 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-config-data\") pod \"57b51201-4e02-4b29-843d-6bfe84247cd8\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.424508 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-sg-core-conf-yaml\") pod \"57b51201-4e02-4b29-843d-6bfe84247cd8\" (UID: \"57b51201-4e02-4b29-843d-6bfe84247cd8\") " Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.426335 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57b51201-4e02-4b29-843d-6bfe84247cd8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "57b51201-4e02-4b29-843d-6bfe84247cd8" (UID: "57b51201-4e02-4b29-843d-6bfe84247cd8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.427321 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57b51201-4e02-4b29-843d-6bfe84247cd8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "57b51201-4e02-4b29-843d-6bfe84247cd8" (UID: "57b51201-4e02-4b29-843d-6bfe84247cd8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.445381 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-scripts" (OuterVolumeSpecName: "scripts") pod "57b51201-4e02-4b29-843d-6bfe84247cd8" (UID: "57b51201-4e02-4b29-843d-6bfe84247cd8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.448338 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b51201-4e02-4b29-843d-6bfe84247cd8-kube-api-access-ctvl6" (OuterVolumeSpecName: "kube-api-access-ctvl6") pod "57b51201-4e02-4b29-843d-6bfe84247cd8" (UID: "57b51201-4e02-4b29-843d-6bfe84247cd8"). InnerVolumeSpecName "kube-api-access-ctvl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.454677 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "57b51201-4e02-4b29-843d-6bfe84247cd8" (UID: "57b51201-4e02-4b29-843d-6bfe84247cd8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.482373 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.521747 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57b51201-4e02-4b29-843d-6bfe84247cd8" (UID: "57b51201-4e02-4b29-843d-6bfe84247cd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.526720 4931 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b51201-4e02-4b29-843d-6bfe84247cd8-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.526775 4931 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.526785 4931 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/57b51201-4e02-4b29-843d-6bfe84247cd8-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.526795 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.526877 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctvl6\" (UniqueName: \"kubernetes.io/projected/57b51201-4e02-4b29-843d-6bfe84247cd8-kube-api-access-ctvl6\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.526890 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.535379 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7d5fd9f77d-c4gvt" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.559008 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-config-data" (OuterVolumeSpecName: "config-data") pod "57b51201-4e02-4b29-843d-6bfe84247cd8" (UID: "57b51201-4e02-4b29-843d-6bfe84247cd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.598623 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8976d4b6b-jvhlw"] Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.599109 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8976d4b6b-jvhlw" podUID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" containerName="barbican-api-log" containerID="cri-o://972ddf9f7dd4c2d2f4b1631eb8207384e9f00326ad71c1534266ebec8b39a7c9" gracePeriod=30 Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.599932 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8976d4b6b-jvhlw" podUID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" containerName="barbican-api" containerID="cri-o://58f6af2ae126f9a0e10c3789cabc10d578d65da09809eff0eb725fa63005b3f3" gracePeriod=30 Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.635258 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57b51201-4e02-4b29-843d-6bfe84247cd8-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:51 crc kubenswrapper[4931]: I1129 00:19:51.939395 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:19:51 crc kubenswrapper[4931]: E1129 00:19:51.939566 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:19:51 crc kubenswrapper[4931]: E1129 00:19:51.939587 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 29 00:19:51 crc kubenswrapper[4931]: E1129 00:19:51.939633 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift podName:a720d879-8b09-452d-8371-9bf75f28cbde nodeName:}" failed. No retries permitted until 2025-11-29 00:21:53.939616306 +0000 UTC m=+1327.101509538 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift") pod "swift-storage-0" (UID: "a720d879-8b09-452d-8371-9bf75f28cbde") : configmap "swift-ring-files" not found Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.281517 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"57b51201-4e02-4b29-843d-6bfe84247cd8","Type":"ContainerDied","Data":"8252cadc8a68ffd1909427114d2e0c176e5f5bd1ac9b013a606a44d9d74ee88f"} Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.281896 4931 scope.go:117] "RemoveContainer" containerID="39d695aa87a1a566d711b861dccd1c5cd384c808f103dbe32f0c6665e379f6e7" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.282069 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.296426 4931 generic.go:334] "Generic (PLEG): container finished" podID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" containerID="972ddf9f7dd4c2d2f4b1631eb8207384e9f00326ad71c1534266ebec8b39a7c9" exitCode=143 Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.296529 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8976d4b6b-jvhlw" event={"ID":"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8","Type":"ContainerDied","Data":"972ddf9f7dd4c2d2f4b1631eb8207384e9f00326ad71c1534266ebec8b39a7c9"} Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.415919 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.427036 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.455152 4931 scope.go:117] "RemoveContainer" containerID="9d01970592fdce0484783d5f757c978cbd0536d5468f72ad79ddae208fd84d3d" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.455848 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:19:52 crc kubenswrapper[4931]: E1129 00:19:52.456240 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="proxy-httpd" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456257 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="proxy-httpd" Nov 29 00:19:52 crc kubenswrapper[4931]: E1129 00:19:52.456273 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="ceilometer-notification-agent" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456279 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="ceilometer-notification-agent" Nov 29 00:19:52 crc kubenswrapper[4931]: E1129 00:19:52.456289 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="sg-core" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456295 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="sg-core" Nov 29 00:19:52 crc kubenswrapper[4931]: E1129 00:19:52.456311 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8826b2d2-afb8-449c-8b42-c3d8e50cfe59" containerName="horizon" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456317 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="8826b2d2-afb8-449c-8b42-c3d8e50cfe59" containerName="horizon" Nov 29 00:19:52 crc kubenswrapper[4931]: E1129 00:19:52.456328 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="653aa9fc-902d-4186-812f-168b8a822c89" containerName="horizon" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456333 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="653aa9fc-902d-4186-812f-168b8a822c89" containerName="horizon" Nov 29 00:19:52 crc kubenswrapper[4931]: E1129 00:19:52.456342 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="653aa9fc-902d-4186-812f-168b8a822c89" containerName="horizon-log" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456347 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="653aa9fc-902d-4186-812f-168b8a822c89" containerName="horizon-log" Nov 29 00:19:52 crc kubenswrapper[4931]: E1129 00:19:52.456358 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8826b2d2-afb8-449c-8b42-c3d8e50cfe59" containerName="horizon-log" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456363 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="8826b2d2-afb8-449c-8b42-c3d8e50cfe59" containerName="horizon-log" Nov 29 00:19:52 crc kubenswrapper[4931]: E1129 00:19:52.456381 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="ceilometer-central-agent" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456388 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="ceilometer-central-agent" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456547 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="ceilometer-central-agent" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456556 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="proxy-httpd" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456567 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="653aa9fc-902d-4186-812f-168b8a822c89" containerName="horizon-log" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456576 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="8826b2d2-afb8-449c-8b42-c3d8e50cfe59" containerName="horizon" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456583 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="ceilometer-notification-agent" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456593 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="653aa9fc-902d-4186-812f-168b8a822c89" containerName="horizon" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456605 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" containerName="sg-core" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.456616 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="8826b2d2-afb8-449c-8b42-c3d8e50cfe59" containerName="horizon-log" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.458386 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.460918 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.461271 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.469304 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.524816 4931 scope.go:117] "RemoveContainer" containerID="8e97f61f77404e3bbfb6af14a2ed78fa4f9311b5b8a206712ad11e9b1b2cb0b4" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.550876 4931 scope.go:117] "RemoveContainer" containerID="bdedfb8b573d44dcd2880bd5f7719721cac9337ba0fe28318bdbf0e5b580ad90" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.551672 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-config-data\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.551748 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-scripts\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.551797 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.551872 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f301267-c058-4269-b1b8-80e81c365ab0-log-httpd\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.551967 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpg5b\" (UniqueName: \"kubernetes.io/projected/4f301267-c058-4269-b1b8-80e81c365ab0-kube-api-access-kpg5b\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.552017 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f301267-c058-4269-b1b8-80e81c365ab0-run-httpd\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.552049 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.654425 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpg5b\" (UniqueName: \"kubernetes.io/projected/4f301267-c058-4269-b1b8-80e81c365ab0-kube-api-access-kpg5b\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.655602 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f301267-c058-4269-b1b8-80e81c365ab0-run-httpd\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.655735 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.656270 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f301267-c058-4269-b1b8-80e81c365ab0-run-httpd\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.656293 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-config-data\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.656481 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-scripts\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.656616 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.656910 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f301267-c058-4269-b1b8-80e81c365ab0-log-httpd\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.658992 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f301267-c058-4269-b1b8-80e81c365ab0-log-httpd\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.664446 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.665518 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.668497 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-scripts\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.674100 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-config-data\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.689611 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpg5b\" (UniqueName: \"kubernetes.io/projected/4f301267-c058-4269-b1b8-80e81c365ab0-kube-api-access-kpg5b\") pod \"ceilometer-0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.786470 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:19:52 crc kubenswrapper[4931]: I1129 00:19:52.884169 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.065493 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-ovndb-tls-certs\") pod \"f7044c78-4dc6-4934-8d91-e8735d23be54\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.065610 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-config\") pod \"f7044c78-4dc6-4934-8d91-e8735d23be54\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.065742 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sqww\" (UniqueName: \"kubernetes.io/projected/f7044c78-4dc6-4934-8d91-e8735d23be54-kube-api-access-2sqww\") pod \"f7044c78-4dc6-4934-8d91-e8735d23be54\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.065791 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-combined-ca-bundle\") pod \"f7044c78-4dc6-4934-8d91-e8735d23be54\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.065847 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-httpd-config\") pod \"f7044c78-4dc6-4934-8d91-e8735d23be54\" (UID: \"f7044c78-4dc6-4934-8d91-e8735d23be54\") " Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.071791 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "f7044c78-4dc6-4934-8d91-e8735d23be54" (UID: "f7044c78-4dc6-4934-8d91-e8735d23be54"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.073215 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7044c78-4dc6-4934-8d91-e8735d23be54-kube-api-access-2sqww" (OuterVolumeSpecName: "kube-api-access-2sqww") pod "f7044c78-4dc6-4934-8d91-e8735d23be54" (UID: "f7044c78-4dc6-4934-8d91-e8735d23be54"). InnerVolumeSpecName "kube-api-access-2sqww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.133773 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7044c78-4dc6-4934-8d91-e8735d23be54" (UID: "f7044c78-4dc6-4934-8d91-e8735d23be54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.136403 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "f7044c78-4dc6-4934-8d91-e8735d23be54" (UID: "f7044c78-4dc6-4934-8d91-e8735d23be54"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.141572 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-config" (OuterVolumeSpecName: "config") pod "f7044c78-4dc6-4934-8d91-e8735d23be54" (UID: "f7044c78-4dc6-4934-8d91-e8735d23be54"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.168210 4931 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.168242 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.168254 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sqww\" (UniqueName: \"kubernetes.io/projected/f7044c78-4dc6-4934-8d91-e8735d23be54-kube-api-access-2sqww\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.168264 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.168274 4931 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f7044c78-4dc6-4934-8d91-e8735d23be54-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.223037 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57b51201-4e02-4b29-843d-6bfe84247cd8" path="/var/lib/kubelet/pods/57b51201-4e02-4b29-843d-6bfe84247cd8/volumes" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.309203 4931 generic.go:334] "Generic (PLEG): container finished" podID="b67e3b5f-e942-4719-ae24-bd98ae698859" containerID="76dfdd0136931cb4a46398d2ae3d1e90ecc3611e19ee01719564dcc2523fb0b9" exitCode=0 Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.310239 4931 generic.go:334] "Generic (PLEG): container finished" podID="b67e3b5f-e942-4719-ae24-bd98ae698859" containerID="fc9ffbd2c9acb232861dc0bdfce3b796cdfed7bada972b6b4068b8812a50ce69" exitCode=0 Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.309261 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b67e3b5f-e942-4719-ae24-bd98ae698859","Type":"ContainerDied","Data":"76dfdd0136931cb4a46398d2ae3d1e90ecc3611e19ee01719564dcc2523fb0b9"} Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.310450 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b67e3b5f-e942-4719-ae24-bd98ae698859","Type":"ContainerDied","Data":"fc9ffbd2c9acb232861dc0bdfce3b796cdfed7bada972b6b4068b8812a50ce69"} Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.313125 4931 generic.go:334] "Generic (PLEG): container finished" podID="f7044c78-4dc6-4934-8d91-e8735d23be54" containerID="4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3" exitCode=0 Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.313177 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b8d6898dd-wpzld" event={"ID":"f7044c78-4dc6-4934-8d91-e8735d23be54","Type":"ContainerDied","Data":"4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3"} Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.313208 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b8d6898dd-wpzld" event={"ID":"f7044c78-4dc6-4934-8d91-e8735d23be54","Type":"ContainerDied","Data":"5e8a0a30fa98196d903e4530a4314cdeb7a7c0846a39ed5a04d32e450b610830"} Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.313228 4931 scope.go:117] "RemoveContainer" containerID="7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.313232 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b8d6898dd-wpzld" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.338474 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.346358 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b8d6898dd-wpzld"] Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.350458 4931 scope.go:117] "RemoveContainer" containerID="4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.358910 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b8d6898dd-wpzld"] Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.395002 4931 scope.go:117] "RemoveContainer" containerID="7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122" Nov 29 00:19:53 crc kubenswrapper[4931]: E1129 00:19:53.395991 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122\": container with ID starting with 7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122 not found: ID does not exist" containerID="7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.396038 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122"} err="failed to get container status \"7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122\": rpc error: code = NotFound desc = could not find container \"7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122\": container with ID starting with 7378081992cda7305abfaf7e128151bf3bc2caca1085dc294671c6c817abb122 not found: ID does not exist" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.396065 4931 scope.go:117] "RemoveContainer" containerID="4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3" Nov 29 00:19:53 crc kubenswrapper[4931]: E1129 00:19:53.396448 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3\": container with ID starting with 4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3 not found: ID does not exist" containerID="4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.396544 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3"} err="failed to get container status \"4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3\": rpc error: code = NotFound desc = could not find container \"4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3\": container with ID starting with 4d14f22675d2062dd7da4c46e6d7f34887093ef6958dea53cd83ed8781b433a3 not found: ID does not exist" Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.735198 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:19:53 crc kubenswrapper[4931]: I1129 00:19:53.735511 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:19:54 crc kubenswrapper[4931]: I1129 00:19:54.329335 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f301267-c058-4269-b1b8-80e81c365ab0","Type":"ContainerStarted","Data":"3f1c49a83908f4d4f53fd573635105dcbe347dfa65b535267f38ce1e0096c039"} Nov 29 00:19:54 crc kubenswrapper[4931]: I1129 00:19:54.794570 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 29 00:19:55 crc kubenswrapper[4931]: I1129 00:19:55.033158 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8976d4b6b-jvhlw" podUID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": read tcp 10.217.0.2:57930->10.217.0.157:9311: read: connection reset by peer" Nov 29 00:19:55 crc kubenswrapper[4931]: I1129 00:19:55.033160 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8976d4b6b-jvhlw" podUID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": read tcp 10.217.0.2:57942->10.217.0.157:9311: read: connection reset by peer" Nov 29 00:19:55 crc kubenswrapper[4931]: I1129 00:19:55.227450 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7044c78-4dc6-4934-8d91-e8735d23be54" path="/var/lib/kubelet/pods/f7044c78-4dc6-4934-8d91-e8735d23be54/volumes" Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.353997 4931 generic.go:334] "Generic (PLEG): container finished" podID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" containerID="58f6af2ae126f9a0e10c3789cabc10d578d65da09809eff0eb725fa63005b3f3" exitCode=0 Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.354071 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8976d4b6b-jvhlw" event={"ID":"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8","Type":"ContainerDied","Data":"58f6af2ae126f9a0e10c3789cabc10d578d65da09809eff0eb725fa63005b3f3"} Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.774540 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.846654 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-config-data\") pod \"b67e3b5f-e942-4719-ae24-bd98ae698859\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.847180 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-config-data-custom\") pod \"b67e3b5f-e942-4719-ae24-bd98ae698859\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.847257 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b67e3b5f-e942-4719-ae24-bd98ae698859-etc-machine-id\") pod \"b67e3b5f-e942-4719-ae24-bd98ae698859\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.847287 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-scripts\") pod \"b67e3b5f-e942-4719-ae24-bd98ae698859\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.847372 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-combined-ca-bundle\") pod \"b67e3b5f-e942-4719-ae24-bd98ae698859\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.847409 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7mj9\" (UniqueName: \"kubernetes.io/projected/b67e3b5f-e942-4719-ae24-bd98ae698859-kube-api-access-q7mj9\") pod \"b67e3b5f-e942-4719-ae24-bd98ae698859\" (UID: \"b67e3b5f-e942-4719-ae24-bd98ae698859\") " Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.848890 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b67e3b5f-e942-4719-ae24-bd98ae698859-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b67e3b5f-e942-4719-ae24-bd98ae698859" (UID: "b67e3b5f-e942-4719-ae24-bd98ae698859"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.861169 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b67e3b5f-e942-4719-ae24-bd98ae698859" (UID: "b67e3b5f-e942-4719-ae24-bd98ae698859"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.867636 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-scripts" (OuterVolumeSpecName: "scripts") pod "b67e3b5f-e942-4719-ae24-bd98ae698859" (UID: "b67e3b5f-e942-4719-ae24-bd98ae698859"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.903096 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b67e3b5f-e942-4719-ae24-bd98ae698859-kube-api-access-q7mj9" (OuterVolumeSpecName: "kube-api-access-q7mj9") pod "b67e3b5f-e942-4719-ae24-bd98ae698859" (UID: "b67e3b5f-e942-4719-ae24-bd98ae698859"). InnerVolumeSpecName "kube-api-access-q7mj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.936081 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.951418 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7mj9\" (UniqueName: \"kubernetes.io/projected/b67e3b5f-e942-4719-ae24-bd98ae698859-kube-api-access-q7mj9\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.951444 4931 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.951487 4931 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b67e3b5f-e942-4719-ae24-bd98ae698859-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:56 crc kubenswrapper[4931]: I1129 00:19:56.951495 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.000040 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-config-data" (OuterVolumeSpecName: "config-data") pod "b67e3b5f-e942-4719-ae24-bd98ae698859" (UID: "b67e3b5f-e942-4719-ae24-bd98ae698859"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.002999 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6b5869645b-zlntx" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.014118 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.074500 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b67e3b5f-e942-4719-ae24-bd98ae698859" (UID: "b67e3b5f-e942-4719-ae24-bd98ae698859"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.087374 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-config-data\") pod \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.087671 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-logs\") pod \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.087779 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44tpl\" (UniqueName: \"kubernetes.io/projected/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-kube-api-access-44tpl\") pod \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.088025 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-config-data-custom\") pod \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.088172 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-combined-ca-bundle\") pod \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\" (UID: \"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8\") " Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.088698 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.088952 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b67e3b5f-e942-4719-ae24-bd98ae698859-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.090058 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-logs" (OuterVolumeSpecName: "logs") pod "bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" (UID: "bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.098347 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" (UID: "bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.106035 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-kube-api-access-44tpl" (OuterVolumeSpecName: "kube-api-access-44tpl") pod "bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" (UID: "bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8"). InnerVolumeSpecName "kube-api-access-44tpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.121947 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" (UID: "bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.166154 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-config-data" (OuterVolumeSpecName: "config-data") pod "bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" (UID: "bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.191348 4931 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.191378 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.191386 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.191396 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.191406 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44tpl\" (UniqueName: \"kubernetes.io/projected/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8-kube-api-access-44tpl\") on node \"crc\" DevicePath \"\"" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.363836 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b67e3b5f-e942-4719-ae24-bd98ae698859","Type":"ContainerDied","Data":"e6f9cc02dfc02e085e5d790594bcb6e8cddfbfd587001f05eaa649eeba198b85"} Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.363897 4931 scope.go:117] "RemoveContainer" containerID="76dfdd0136931cb4a46398d2ae3d1e90ecc3611e19ee01719564dcc2523fb0b9" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.363939 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.371246 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8976d4b6b-jvhlw" event={"ID":"bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8","Type":"ContainerDied","Data":"68d02929f2e8a85d65ec9d44ccb6d2166597a751c2a86307224a00d15a88124d"} Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.371273 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8976d4b6b-jvhlw" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.392674 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.400754 4931 scope.go:117] "RemoveContainer" containerID="fc9ffbd2c9acb232861dc0bdfce3b796cdfed7bada972b6b4068b8812a50ce69" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.404876 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.419489 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8976d4b6b-jvhlw"] Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.431222 4931 scope.go:117] "RemoveContainer" containerID="58f6af2ae126f9a0e10c3789cabc10d578d65da09809eff0eb725fa63005b3f3" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.431682 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-8976d4b6b-jvhlw"] Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.440165 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 00:19:57 crc kubenswrapper[4931]: E1129 00:19:57.442986 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b67e3b5f-e942-4719-ae24-bd98ae698859" containerName="cinder-scheduler" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.443031 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b67e3b5f-e942-4719-ae24-bd98ae698859" containerName="cinder-scheduler" Nov 29 00:19:57 crc kubenswrapper[4931]: E1129 00:19:57.443066 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" containerName="barbican-api" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.443072 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" containerName="barbican-api" Nov 29 00:19:57 crc kubenswrapper[4931]: E1129 00:19:57.443084 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" containerName="barbican-api-log" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.443093 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" containerName="barbican-api-log" Nov 29 00:19:57 crc kubenswrapper[4931]: E1129 00:19:57.443117 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7044c78-4dc6-4934-8d91-e8735d23be54" containerName="neutron-httpd" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.443123 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7044c78-4dc6-4934-8d91-e8735d23be54" containerName="neutron-httpd" Nov 29 00:19:57 crc kubenswrapper[4931]: E1129 00:19:57.443141 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7044c78-4dc6-4934-8d91-e8735d23be54" containerName="neutron-api" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.443147 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7044c78-4dc6-4934-8d91-e8735d23be54" containerName="neutron-api" Nov 29 00:19:57 crc kubenswrapper[4931]: E1129 00:19:57.443174 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b67e3b5f-e942-4719-ae24-bd98ae698859" containerName="probe" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.443184 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b67e3b5f-e942-4719-ae24-bd98ae698859" containerName="probe" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.443483 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b67e3b5f-e942-4719-ae24-bd98ae698859" containerName="cinder-scheduler" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.443501 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7044c78-4dc6-4934-8d91-e8735d23be54" containerName="neutron-httpd" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.443511 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b67e3b5f-e942-4719-ae24-bd98ae698859" containerName="probe" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.443537 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7044c78-4dc6-4934-8d91-e8735d23be54" containerName="neutron-api" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.443551 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" containerName="barbican-api-log" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.443563 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" containerName="barbican-api" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.444740 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.449758 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.450563 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.464373 4931 scope.go:117] "RemoveContainer" containerID="972ddf9f7dd4c2d2f4b1631eb8207384e9f00326ad71c1534266ebec8b39a7c9" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.496751 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e9c000-a994-4023-a6a1-eaf84b161005-config-data\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.496903 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06e9c000-a994-4023-a6a1-eaf84b161005-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.497036 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf9z5\" (UniqueName: \"kubernetes.io/projected/06e9c000-a994-4023-a6a1-eaf84b161005-kube-api-access-jf9z5\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.497141 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06e9c000-a994-4023-a6a1-eaf84b161005-scripts\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.497222 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e9c000-a994-4023-a6a1-eaf84b161005-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.497317 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06e9c000-a994-4023-a6a1-eaf84b161005-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.596669 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-84d76c555c-shgd6" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.599096 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf9z5\" (UniqueName: \"kubernetes.io/projected/06e9c000-a994-4023-a6a1-eaf84b161005-kube-api-access-jf9z5\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.599221 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06e9c000-a994-4023-a6a1-eaf84b161005-scripts\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.599302 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e9c000-a994-4023-a6a1-eaf84b161005-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.599410 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06e9c000-a994-4023-a6a1-eaf84b161005-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.599519 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e9c000-a994-4023-a6a1-eaf84b161005-config-data\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.599589 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06e9c000-a994-4023-a6a1-eaf84b161005-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.599950 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06e9c000-a994-4023-a6a1-eaf84b161005-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.603594 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06e9c000-a994-4023-a6a1-eaf84b161005-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.603683 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e9c000-a994-4023-a6a1-eaf84b161005-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.605459 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06e9c000-a994-4023-a6a1-eaf84b161005-scripts\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.614636 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e9c000-a994-4023-a6a1-eaf84b161005-config-data\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.639374 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf9z5\" (UniqueName: \"kubernetes.io/projected/06e9c000-a994-4023-a6a1-eaf84b161005-kube-api-access-jf9z5\") pod \"cinder-scheduler-0\" (UID: \"06e9c000-a994-4023-a6a1-eaf84b161005\") " pod="openstack/cinder-scheduler-0" Nov 29 00:19:57 crc kubenswrapper[4931]: I1129 00:19:57.769455 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 29 00:19:58 crc kubenswrapper[4931]: I1129 00:19:58.255788 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 29 00:19:58 crc kubenswrapper[4931]: I1129 00:19:58.381124 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f301267-c058-4269-b1b8-80e81c365ab0","Type":"ContainerStarted","Data":"eef131d6df4d80b9ed2e0b39e6d8faf2c63b90c57e985394e51bfdf995b6ca54"} Nov 29 00:19:58 crc kubenswrapper[4931]: I1129 00:19:58.385866 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06e9c000-a994-4023-a6a1-eaf84b161005","Type":"ContainerStarted","Data":"a5e455672405d334a15a62eb65b101964ffa171bd9a4ee2805e519e6df4e1e8b"} Nov 29 00:19:59 crc kubenswrapper[4931]: I1129 00:19:59.242733 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b67e3b5f-e942-4719-ae24-bd98ae698859" path="/var/lib/kubelet/pods/b67e3b5f-e942-4719-ae24-bd98ae698859/volumes" Nov 29 00:19:59 crc kubenswrapper[4931]: I1129 00:19:59.244463 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8" path="/var/lib/kubelet/pods/bce5bf07-2d1e-4cea-9594-2a0ff2deb7e8/volumes" Nov 29 00:19:59 crc kubenswrapper[4931]: I1129 00:19:59.398353 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f301267-c058-4269-b1b8-80e81c365ab0","Type":"ContainerStarted","Data":"a352a21851f31cfebe861cc2f13c23b28ffbdcf912cdf29a7574a421334a9a89"} Nov 29 00:19:59 crc kubenswrapper[4931]: I1129 00:19:59.400258 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06e9c000-a994-4023-a6a1-eaf84b161005","Type":"ContainerStarted","Data":"baf3542f1771f9e661ff93c0c77f4a7d8455147a0af6798e7f21370e9b22131a"} Nov 29 00:20:00 crc kubenswrapper[4931]: I1129 00:20:00.411693 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06e9c000-a994-4023-a6a1-eaf84b161005","Type":"ContainerStarted","Data":"3c3114ee3f6b0e337fe6a8b694424bc3107009a35202f235919956cba8480c83"} Nov 29 00:20:00 crc kubenswrapper[4931]: I1129 00:20:00.434863 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.434846955 podStartE2EDuration="3.434846955s" podCreationTimestamp="2025-11-29 00:19:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:20:00.43013898 +0000 UTC m=+1213.592032222" watchObservedRunningTime="2025-11-29 00:20:00.434846955 +0000 UTC m=+1213.596740187" Nov 29 00:20:01 crc kubenswrapper[4931]: I1129 00:20:01.344927 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-765c4d856b-6tkcn" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Nov 29 00:20:01 crc kubenswrapper[4931]: I1129 00:20:01.423158 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f301267-c058-4269-b1b8-80e81c365ab0","Type":"ContainerStarted","Data":"ffe62b36cd8545febfa8017c8b2b8891b63c4ac8f1b08f7c2ab3acd4f079107e"} Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.014541 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.021228 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.024460 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.024464 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-9jwkw" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.024597 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.029965 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.175273 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c9276f25-9d5a-4953-9145-4da299973ff4-openstack-config-secret\") pod \"openstackclient\" (UID: \"c9276f25-9d5a-4953-9145-4da299973ff4\") " pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.175639 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8t7t\" (UniqueName: \"kubernetes.io/projected/c9276f25-9d5a-4953-9145-4da299973ff4-kube-api-access-m8t7t\") pod \"openstackclient\" (UID: \"c9276f25-9d5a-4953-9145-4da299973ff4\") " pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.175903 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9276f25-9d5a-4953-9145-4da299973ff4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c9276f25-9d5a-4953-9145-4da299973ff4\") " pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.175944 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c9276f25-9d5a-4953-9145-4da299973ff4-openstack-config\") pod \"openstackclient\" (UID: \"c9276f25-9d5a-4953-9145-4da299973ff4\") " pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.277198 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c9276f25-9d5a-4953-9145-4da299973ff4-openstack-config\") pod \"openstackclient\" (UID: \"c9276f25-9d5a-4953-9145-4da299973ff4\") " pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.277232 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9276f25-9d5a-4953-9145-4da299973ff4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c9276f25-9d5a-4953-9145-4da299973ff4\") " pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.277274 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c9276f25-9d5a-4953-9145-4da299973ff4-openstack-config-secret\") pod \"openstackclient\" (UID: \"c9276f25-9d5a-4953-9145-4da299973ff4\") " pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.277345 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8t7t\" (UniqueName: \"kubernetes.io/projected/c9276f25-9d5a-4953-9145-4da299973ff4-kube-api-access-m8t7t\") pod \"openstackclient\" (UID: \"c9276f25-9d5a-4953-9145-4da299973ff4\") " pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.278377 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c9276f25-9d5a-4953-9145-4da299973ff4-openstack-config\") pod \"openstackclient\" (UID: \"c9276f25-9d5a-4953-9145-4da299973ff4\") " pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.287273 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c9276f25-9d5a-4953-9145-4da299973ff4-openstack-config-secret\") pod \"openstackclient\" (UID: \"c9276f25-9d5a-4953-9145-4da299973ff4\") " pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.294913 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9276f25-9d5a-4953-9145-4da299973ff4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c9276f25-9d5a-4953-9145-4da299973ff4\") " pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.295335 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8t7t\" (UniqueName: \"kubernetes.io/projected/c9276f25-9d5a-4953-9145-4da299973ff4-kube-api-access-m8t7t\") pod \"openstackclient\" (UID: \"c9276f25-9d5a-4953-9145-4da299973ff4\") " pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.341900 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.770426 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 29 00:20:02 crc kubenswrapper[4931]: I1129 00:20:02.817478 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 29 00:20:03 crc kubenswrapper[4931]: I1129 00:20:03.439201 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c9276f25-9d5a-4953-9145-4da299973ff4","Type":"ContainerStarted","Data":"060994830c0f1d011c8655f91f5d0e908a1860d173470edfd2e933811f698b2f"} Nov 29 00:20:03 crc kubenswrapper[4931]: I1129 00:20:03.441679 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f301267-c058-4269-b1b8-80e81c365ab0","Type":"ContainerStarted","Data":"5860656742da10506ce709797e6048eb8aac9d44678aff77fe87164674143d51"} Nov 29 00:20:03 crc kubenswrapper[4931]: I1129 00:20:03.441805 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 00:20:03 crc kubenswrapper[4931]: I1129 00:20:03.480228 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.011537623 podStartE2EDuration="11.480207326s" podCreationTimestamp="2025-11-29 00:19:52 +0000 UTC" firstStartedPulling="2025-11-29 00:19:53.354408594 +0000 UTC m=+1206.516301826" lastFinishedPulling="2025-11-29 00:20:02.823078307 +0000 UTC m=+1215.984971529" observedRunningTime="2025-11-29 00:20:03.472968188 +0000 UTC m=+1216.634861420" watchObservedRunningTime="2025-11-29 00:20:03.480207326 +0000 UTC m=+1216.642100558" Nov 29 00:20:03 crc kubenswrapper[4931]: I1129 00:20:03.562356 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.003987 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-764cb44577-7q67f"] Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.005851 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.008555 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.009665 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.010566 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.035635 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-764cb44577-7q67f"] Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.109968 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/550e16f7-f101-4a91-8e2f-45af3430d6b9-log-httpd\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.110038 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/550e16f7-f101-4a91-8e2f-45af3430d6b9-internal-tls-certs\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.110086 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/550e16f7-f101-4a91-8e2f-45af3430d6b9-combined-ca-bundle\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.110110 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/550e16f7-f101-4a91-8e2f-45af3430d6b9-public-tls-certs\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.110252 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/550e16f7-f101-4a91-8e2f-45af3430d6b9-config-data\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.110369 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.110424 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/550e16f7-f101-4a91-8e2f-45af3430d6b9-run-httpd\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.110454 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8tpl\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-kube-api-access-j8tpl\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.212413 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/550e16f7-f101-4a91-8e2f-45af3430d6b9-log-httpd\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.212452 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/550e16f7-f101-4a91-8e2f-45af3430d6b9-internal-tls-certs\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.212527 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/550e16f7-f101-4a91-8e2f-45af3430d6b9-combined-ca-bundle\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.212551 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/550e16f7-f101-4a91-8e2f-45af3430d6b9-public-tls-certs\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.212659 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/550e16f7-f101-4a91-8e2f-45af3430d6b9-config-data\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.213495 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/550e16f7-f101-4a91-8e2f-45af3430d6b9-log-httpd\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.213841 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.213898 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/550e16f7-f101-4a91-8e2f-45af3430d6b9-run-httpd\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.213963 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8tpl\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-kube-api-access-j8tpl\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: E1129 00:20:04.216571 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:20:04 crc kubenswrapper[4931]: E1129 00:20:04.216601 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-764cb44577-7q67f: configmap "swift-ring-files" not found Nov 29 00:20:04 crc kubenswrapper[4931]: E1129 00:20:04.216658 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift podName:550e16f7-f101-4a91-8e2f-45af3430d6b9 nodeName:}" failed. No retries permitted until 2025-11-29 00:20:04.716638225 +0000 UTC m=+1217.878531537 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift") pod "swift-proxy-764cb44577-7q67f" (UID: "550e16f7-f101-4a91-8e2f-45af3430d6b9") : configmap "swift-ring-files" not found Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.218252 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/550e16f7-f101-4a91-8e2f-45af3430d6b9-run-httpd\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.220317 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/550e16f7-f101-4a91-8e2f-45af3430d6b9-config-data\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.221563 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/550e16f7-f101-4a91-8e2f-45af3430d6b9-public-tls-certs\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.224495 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/550e16f7-f101-4a91-8e2f-45af3430d6b9-internal-tls-certs\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.224940 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/550e16f7-f101-4a91-8e2f-45af3430d6b9-combined-ca-bundle\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.236247 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8tpl\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-kube-api-access-j8tpl\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: I1129 00:20:04.726948 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:04 crc kubenswrapper[4931]: E1129 00:20:04.727170 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:20:04 crc kubenswrapper[4931]: E1129 00:20:04.727545 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-764cb44577-7q67f: configmap "swift-ring-files" not found Nov 29 00:20:04 crc kubenswrapper[4931]: E1129 00:20:04.727612 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift podName:550e16f7-f101-4a91-8e2f-45af3430d6b9 nodeName:}" failed. No retries permitted until 2025-11-29 00:20:05.727589383 +0000 UTC m=+1218.889482615 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift") pod "swift-proxy-764cb44577-7q67f" (UID: "550e16f7-f101-4a91-8e2f-45af3430d6b9") : configmap "swift-ring-files" not found Nov 29 00:20:05 crc kubenswrapper[4931]: I1129 00:20:05.464371 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="ceilometer-central-agent" containerID="cri-o://eef131d6df4d80b9ed2e0b39e6d8faf2c63b90c57e985394e51bfdf995b6ca54" gracePeriod=30 Nov 29 00:20:05 crc kubenswrapper[4931]: I1129 00:20:05.464398 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="sg-core" containerID="cri-o://ffe62b36cd8545febfa8017c8b2b8891b63c4ac8f1b08f7c2ab3acd4f079107e" gracePeriod=30 Nov 29 00:20:05 crc kubenswrapper[4931]: I1129 00:20:05.464418 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="ceilometer-notification-agent" containerID="cri-o://a352a21851f31cfebe861cc2f13c23b28ffbdcf912cdf29a7574a421334a9a89" gracePeriod=30 Nov 29 00:20:05 crc kubenswrapper[4931]: I1129 00:20:05.464477 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="proxy-httpd" containerID="cri-o://5860656742da10506ce709797e6048eb8aac9d44678aff77fe87164674143d51" gracePeriod=30 Nov 29 00:20:05 crc kubenswrapper[4931]: I1129 00:20:05.744036 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:05 crc kubenswrapper[4931]: E1129 00:20:05.744259 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:20:05 crc kubenswrapper[4931]: E1129 00:20:05.744272 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-764cb44577-7q67f: configmap "swift-ring-files" not found Nov 29 00:20:05 crc kubenswrapper[4931]: E1129 00:20:05.744313 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift podName:550e16f7-f101-4a91-8e2f-45af3430d6b9 nodeName:}" failed. No retries permitted until 2025-11-29 00:20:07.744300099 +0000 UTC m=+1220.906193331 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift") pod "swift-proxy-764cb44577-7q67f" (UID: "550e16f7-f101-4a91-8e2f-45af3430d6b9") : configmap "swift-ring-files" not found Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.475140 4931 generic.go:334] "Generic (PLEG): container finished" podID="4f301267-c058-4269-b1b8-80e81c365ab0" containerID="5860656742da10506ce709797e6048eb8aac9d44678aff77fe87164674143d51" exitCode=0 Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.475382 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f301267-c058-4269-b1b8-80e81c365ab0","Type":"ContainerDied","Data":"5860656742da10506ce709797e6048eb8aac9d44678aff77fe87164674143d51"} Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.475471 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f301267-c058-4269-b1b8-80e81c365ab0","Type":"ContainerDied","Data":"ffe62b36cd8545febfa8017c8b2b8891b63c4ac8f1b08f7c2ab3acd4f079107e"} Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.475412 4931 generic.go:334] "Generic (PLEG): container finished" podID="4f301267-c058-4269-b1b8-80e81c365ab0" containerID="ffe62b36cd8545febfa8017c8b2b8891b63c4ac8f1b08f7c2ab3acd4f079107e" exitCode=2 Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.475505 4931 generic.go:334] "Generic (PLEG): container finished" podID="4f301267-c058-4269-b1b8-80e81c365ab0" containerID="a352a21851f31cfebe861cc2f13c23b28ffbdcf912cdf29a7574a421334a9a89" exitCode=0 Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.475551 4931 generic.go:334] "Generic (PLEG): container finished" podID="4f301267-c058-4269-b1b8-80e81c365ab0" containerID="eef131d6df4d80b9ed2e0b39e6d8faf2c63b90c57e985394e51bfdf995b6ca54" exitCode=0 Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.475569 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f301267-c058-4269-b1b8-80e81c365ab0","Type":"ContainerDied","Data":"a352a21851f31cfebe861cc2f13c23b28ffbdcf912cdf29a7574a421334a9a89"} Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.475601 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f301267-c058-4269-b1b8-80e81c365ab0","Type":"ContainerDied","Data":"eef131d6df4d80b9ed2e0b39e6d8faf2c63b90c57e985394e51bfdf995b6ca54"} Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.832670 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.963569 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-scripts\") pod \"4f301267-c058-4269-b1b8-80e81c365ab0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.963676 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f301267-c058-4269-b1b8-80e81c365ab0-log-httpd\") pod \"4f301267-c058-4269-b1b8-80e81c365ab0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.963703 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-config-data\") pod \"4f301267-c058-4269-b1b8-80e81c365ab0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.963853 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpg5b\" (UniqueName: \"kubernetes.io/projected/4f301267-c058-4269-b1b8-80e81c365ab0-kube-api-access-kpg5b\") pod \"4f301267-c058-4269-b1b8-80e81c365ab0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.963912 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-sg-core-conf-yaml\") pod \"4f301267-c058-4269-b1b8-80e81c365ab0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.963936 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f301267-c058-4269-b1b8-80e81c365ab0-run-httpd\") pod \"4f301267-c058-4269-b1b8-80e81c365ab0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.964009 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-combined-ca-bundle\") pod \"4f301267-c058-4269-b1b8-80e81c365ab0\" (UID: \"4f301267-c058-4269-b1b8-80e81c365ab0\") " Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.964273 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f301267-c058-4269-b1b8-80e81c365ab0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4f301267-c058-4269-b1b8-80e81c365ab0" (UID: "4f301267-c058-4269-b1b8-80e81c365ab0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.964510 4931 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f301267-c058-4269-b1b8-80e81c365ab0-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.965705 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f301267-c058-4269-b1b8-80e81c365ab0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4f301267-c058-4269-b1b8-80e81c365ab0" (UID: "4f301267-c058-4269-b1b8-80e81c365ab0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.970601 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-scripts" (OuterVolumeSpecName: "scripts") pod "4f301267-c058-4269-b1b8-80e81c365ab0" (UID: "4f301267-c058-4269-b1b8-80e81c365ab0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.978019 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f301267-c058-4269-b1b8-80e81c365ab0-kube-api-access-kpg5b" (OuterVolumeSpecName: "kube-api-access-kpg5b") pod "4f301267-c058-4269-b1b8-80e81c365ab0" (UID: "4f301267-c058-4269-b1b8-80e81c365ab0"). InnerVolumeSpecName "kube-api-access-kpg5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:20:06 crc kubenswrapper[4931]: I1129 00:20:06.996720 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4f301267-c058-4269-b1b8-80e81c365ab0" (UID: "4f301267-c058-4269-b1b8-80e81c365ab0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.069510 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpg5b\" (UniqueName: \"kubernetes.io/projected/4f301267-c058-4269-b1b8-80e81c365ab0-kube-api-access-kpg5b\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.069958 4931 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.069971 4931 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f301267-c058-4269-b1b8-80e81c365ab0-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.069981 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.072794 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f301267-c058-4269-b1b8-80e81c365ab0" (UID: "4f301267-c058-4269-b1b8-80e81c365ab0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.080369 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-config-data" (OuterVolumeSpecName: "config-data") pod "4f301267-c058-4269-b1b8-80e81c365ab0" (UID: "4f301267-c058-4269-b1b8-80e81c365ab0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.172126 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.172163 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f301267-c058-4269-b1b8-80e81c365ab0-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.490473 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f301267-c058-4269-b1b8-80e81c365ab0","Type":"ContainerDied","Data":"3f1c49a83908f4d4f53fd573635105dcbe347dfa65b535267f38ce1e0096c039"} Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.490542 4931 scope.go:117] "RemoveContainer" containerID="5860656742da10506ce709797e6048eb8aac9d44678aff77fe87164674143d51" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.490703 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.525922 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.545549 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.563855 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:07 crc kubenswrapper[4931]: E1129 00:20:07.564269 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="ceilometer-notification-agent" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.564282 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="ceilometer-notification-agent" Nov 29 00:20:07 crc kubenswrapper[4931]: E1129 00:20:07.564290 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="sg-core" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.564296 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="sg-core" Nov 29 00:20:07 crc kubenswrapper[4931]: E1129 00:20:07.564310 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="ceilometer-central-agent" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.564317 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="ceilometer-central-agent" Nov 29 00:20:07 crc kubenswrapper[4931]: E1129 00:20:07.564329 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="proxy-httpd" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.564335 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="proxy-httpd" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.564523 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="sg-core" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.564538 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="proxy-httpd" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.564548 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="ceilometer-central-agent" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.564559 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" containerName="ceilometer-notification-agent" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.567506 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.569566 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.569906 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.576591 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.678791 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/156abbdb-244c-407b-94b1-80c8353222f5-run-httpd\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.678870 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.678909 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/156abbdb-244c-407b-94b1-80c8353222f5-log-httpd\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.678970 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-config-data\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.678991 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.679005 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-scripts\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.679030 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvr58\" (UniqueName: \"kubernetes.io/projected/156abbdb-244c-407b-94b1-80c8353222f5-kube-api-access-pvr58\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.780510 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/156abbdb-244c-407b-94b1-80c8353222f5-log-httpd\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.780616 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-config-data\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.780644 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.780660 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-scripts\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.780692 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvr58\" (UniqueName: \"kubernetes.io/projected/156abbdb-244c-407b-94b1-80c8353222f5-kube-api-access-pvr58\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.780744 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.780846 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/156abbdb-244c-407b-94b1-80c8353222f5-run-httpd\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.780881 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.781122 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/156abbdb-244c-407b-94b1-80c8353222f5-log-httpd\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: E1129 00:20:07.781536 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:20:07 crc kubenswrapper[4931]: E1129 00:20:07.781554 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-764cb44577-7q67f: configmap "swift-ring-files" not found Nov 29 00:20:07 crc kubenswrapper[4931]: E1129 00:20:07.781612 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift podName:550e16f7-f101-4a91-8e2f-45af3430d6b9 nodeName:}" failed. No retries permitted until 2025-11-29 00:20:11.781586932 +0000 UTC m=+1224.943480164 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift") pod "swift-proxy-764cb44577-7q67f" (UID: "550e16f7-f101-4a91-8e2f-45af3430d6b9") : configmap "swift-ring-files" not found Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.782321 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/156abbdb-244c-407b-94b1-80c8353222f5-run-httpd\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.785935 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-config-data\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.794193 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-scripts\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.800622 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.801176 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvr58\" (UniqueName: \"kubernetes.io/projected/156abbdb-244c-407b-94b1-80c8353222f5-kube-api-access-pvr58\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.805498 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " pod="openstack/ceilometer-0" Nov 29 00:20:07 crc kubenswrapper[4931]: I1129 00:20:07.888984 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:20:08 crc kubenswrapper[4931]: I1129 00:20:08.001392 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 29 00:20:09 crc kubenswrapper[4931]: I1129 00:20:09.227999 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f301267-c058-4269-b1b8-80e81c365ab0" path="/var/lib/kubelet/pods/4f301267-c058-4269-b1b8-80e81c365ab0/volumes" Nov 29 00:20:11 crc kubenswrapper[4931]: I1129 00:20:11.344140 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-765c4d856b-6tkcn" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Nov 29 00:20:11 crc kubenswrapper[4931]: I1129 00:20:11.344557 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:20:11 crc kubenswrapper[4931]: I1129 00:20:11.856069 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:11 crc kubenswrapper[4931]: E1129 00:20:11.856214 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:20:11 crc kubenswrapper[4931]: E1129 00:20:11.856506 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-764cb44577-7q67f: configmap "swift-ring-files" not found Nov 29 00:20:11 crc kubenswrapper[4931]: E1129 00:20:11.856567 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift podName:550e16f7-f101-4a91-8e2f-45af3430d6b9 nodeName:}" failed. No retries permitted until 2025-11-29 00:20:19.856548059 +0000 UTC m=+1233.018441291 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift") pod "swift-proxy-764cb44577-7q67f" (UID: "550e16f7-f101-4a91-8e2f-45af3430d6b9") : configmap "swift-ring-files" not found Nov 29 00:20:12 crc kubenswrapper[4931]: I1129 00:20:12.130492 4931 scope.go:117] "RemoveContainer" containerID="ffe62b36cd8545febfa8017c8b2b8891b63c4ac8f1b08f7c2ab3acd4f079107e" Nov 29 00:20:12 crc kubenswrapper[4931]: I1129 00:20:12.192881 4931 scope.go:117] "RemoveContainer" containerID="a352a21851f31cfebe861cc2f13c23b28ffbdcf912cdf29a7574a421334a9a89" Nov 29 00:20:12 crc kubenswrapper[4931]: I1129 00:20:12.355998 4931 scope.go:117] "RemoveContainer" containerID="eef131d6df4d80b9ed2e0b39e6d8faf2c63b90c57e985394e51bfdf995b6ca54" Nov 29 00:20:12 crc kubenswrapper[4931]: I1129 00:20:12.447989 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:12 crc kubenswrapper[4931]: I1129 00:20:12.550264 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c9276f25-9d5a-4953-9145-4da299973ff4","Type":"ContainerStarted","Data":"6aac11de61f42bd95b39b2bd89f4789c650b8bdbe2de3c9b68402e4448398ec6"} Nov 29 00:20:12 crc kubenswrapper[4931]: I1129 00:20:12.571581 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.165479921 podStartE2EDuration="11.571557913s" podCreationTimestamp="2025-11-29 00:20:01 +0000 UTC" firstStartedPulling="2025-11-29 00:20:02.822365517 +0000 UTC m=+1215.984258749" lastFinishedPulling="2025-11-29 00:20:12.228443479 +0000 UTC m=+1225.390336741" observedRunningTime="2025-11-29 00:20:12.565928321 +0000 UTC m=+1225.727821563" watchObservedRunningTime="2025-11-29 00:20:12.571557913 +0000 UTC m=+1225.733451145" Nov 29 00:20:12 crc kubenswrapper[4931]: I1129 00:20:12.670993 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:13 crc kubenswrapper[4931]: I1129 00:20:13.562954 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"156abbdb-244c-407b-94b1-80c8353222f5","Type":"ContainerStarted","Data":"4fbc00259382e453e90415fcc7e096975e8abcd9ffc553bdcb3947360b85da6c"} Nov 29 00:20:13 crc kubenswrapper[4931]: E1129 00:20:13.724302 4931 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/c1b260d2c20a74567159c5838c210588de2801e12de5a8c2c4436763519670c8/diff" to get inode usage: stat /var/lib/containers/storage/overlay/c1b260d2c20a74567159c5838c210588de2801e12de5a8c2c4436763519670c8/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_57b51201-4e02-4b29-843d-6bfe84247cd8/ceilometer-notification-agent/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_57b51201-4e02-4b29-843d-6bfe84247cd8/ceilometer-notification-agent/0.log: no such file or directory Nov 29 00:20:15 crc kubenswrapper[4931]: I1129 00:20:15.580064 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"156abbdb-244c-407b-94b1-80c8353222f5","Type":"ContainerStarted","Data":"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31"} Nov 29 00:20:16 crc kubenswrapper[4931]: W1129 00:20:16.256638 4931 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-3f1c49a83908f4d4f53fd573635105dcbe347dfa65b535267f38ce1e0096c039": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-3f1c49a83908f4d4f53fd573635105dcbe347dfa65b535267f38ce1e0096c039: no such file or directory Nov 29 00:20:16 crc kubenswrapper[4931]: W1129 00:20:16.257082 4931 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-conmon-eef131d6df4d80b9ed2e0b39e6d8faf2c63b90c57e985394e51bfdf995b6ca54.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-conmon-eef131d6df4d80b9ed2e0b39e6d8faf2c63b90c57e985394e51bfdf995b6ca54.scope: no such file or directory Nov 29 00:20:16 crc kubenswrapper[4931]: W1129 00:20:16.257115 4931 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-eef131d6df4d80b9ed2e0b39e6d8faf2c63b90c57e985394e51bfdf995b6ca54.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-eef131d6df4d80b9ed2e0b39e6d8faf2c63b90c57e985394e51bfdf995b6ca54.scope: no such file or directory Nov 29 00:20:16 crc kubenswrapper[4931]: W1129 00:20:16.257703 4931 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-conmon-a352a21851f31cfebe861cc2f13c23b28ffbdcf912cdf29a7574a421334a9a89.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-conmon-a352a21851f31cfebe861cc2f13c23b28ffbdcf912cdf29a7574a421334a9a89.scope: no such file or directory Nov 29 00:20:16 crc kubenswrapper[4931]: W1129 00:20:16.257730 4931 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-a352a21851f31cfebe861cc2f13c23b28ffbdcf912cdf29a7574a421334a9a89.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-a352a21851f31cfebe861cc2f13c23b28ffbdcf912cdf29a7574a421334a9a89.scope: no such file or directory Nov 29 00:20:16 crc kubenswrapper[4931]: W1129 00:20:16.257757 4931 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-conmon-ffe62b36cd8545febfa8017c8b2b8891b63c4ac8f1b08f7c2ab3acd4f079107e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-conmon-ffe62b36cd8545febfa8017c8b2b8891b63c4ac8f1b08f7c2ab3acd4f079107e.scope: no such file or directory Nov 29 00:20:16 crc kubenswrapper[4931]: W1129 00:20:16.257781 4931 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-ffe62b36cd8545febfa8017c8b2b8891b63c4ac8f1b08f7c2ab3acd4f079107e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-ffe62b36cd8545febfa8017c8b2b8891b63c4ac8f1b08f7c2ab3acd4f079107e.scope: no such file or directory Nov 29 00:20:16 crc kubenswrapper[4931]: W1129 00:20:16.260378 4931 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-conmon-5860656742da10506ce709797e6048eb8aac9d44678aff77fe87164674143d51.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-conmon-5860656742da10506ce709797e6048eb8aac9d44678aff77fe87164674143d51.scope: no such file or directory Nov 29 00:20:16 crc kubenswrapper[4931]: W1129 00:20:16.260661 4931 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-5860656742da10506ce709797e6048eb8aac9d44678aff77fe87164674143d51.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice/crio-5860656742da10506ce709797e6048eb8aac9d44678aff77fe87164674143d51.scope: no such file or directory Nov 29 00:20:16 crc kubenswrapper[4931]: E1129 00:20:16.540004 4931 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77f47a1f_3302_4d58_9b94_4ea756cc38ab.slice/crio-conmon-da9b24f68267e8a0f02647614f89605bc826dc9c40f4d64d967492a86d47c0eb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77f47a1f_3302_4d58_9b94_4ea756cc38ab.slice/crio-da9b24f68267e8a0f02647614f89605bc826dc9c40f4d64d967492a86d47c0eb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f301267_c058_4269_b1b8_80e81c365ab0.slice\": RecentStats: unable to find data in memory cache]" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.612124 4931 generic.go:334] "Generic (PLEG): container finished" podID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerID="da9b24f68267e8a0f02647614f89605bc826dc9c40f4d64d967492a86d47c0eb" exitCode=137 Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.612187 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-765c4d856b-6tkcn" event={"ID":"77f47a1f-3302-4d58-9b94-4ea756cc38ab","Type":"ContainerDied","Data":"da9b24f68267e8a0f02647614f89605bc826dc9c40f4d64d967492a86d47c0eb"} Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.614196 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"156abbdb-244c-407b-94b1-80c8353222f5","Type":"ContainerStarted","Data":"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26"} Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.697588 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.781414 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77f47a1f-3302-4d58-9b94-4ea756cc38ab-config-data\") pod \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.781525 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-combined-ca-bundle\") pod \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.781587 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f47a1f-3302-4d58-9b94-4ea756cc38ab-logs\") pod \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.781650 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-horizon-tls-certs\") pod \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.781687 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sljht\" (UniqueName: \"kubernetes.io/projected/77f47a1f-3302-4d58-9b94-4ea756cc38ab-kube-api-access-sljht\") pod \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.781798 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-horizon-secret-key\") pod \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.781901 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77f47a1f-3302-4d58-9b94-4ea756cc38ab-scripts\") pod \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\" (UID: \"77f47a1f-3302-4d58-9b94-4ea756cc38ab\") " Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.787633 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77f47a1f-3302-4d58-9b94-4ea756cc38ab-kube-api-access-sljht" (OuterVolumeSpecName: "kube-api-access-sljht") pod "77f47a1f-3302-4d58-9b94-4ea756cc38ab" (UID: "77f47a1f-3302-4d58-9b94-4ea756cc38ab"). InnerVolumeSpecName "kube-api-access-sljht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.788517 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77f47a1f-3302-4d58-9b94-4ea756cc38ab-logs" (OuterVolumeSpecName: "logs") pod "77f47a1f-3302-4d58-9b94-4ea756cc38ab" (UID: "77f47a1f-3302-4d58-9b94-4ea756cc38ab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.791729 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "77f47a1f-3302-4d58-9b94-4ea756cc38ab" (UID: "77f47a1f-3302-4d58-9b94-4ea756cc38ab"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.815216 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77f47a1f-3302-4d58-9b94-4ea756cc38ab-config-data" (OuterVolumeSpecName: "config-data") pod "77f47a1f-3302-4d58-9b94-4ea756cc38ab" (UID: "77f47a1f-3302-4d58-9b94-4ea756cc38ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.829525 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77f47a1f-3302-4d58-9b94-4ea756cc38ab-scripts" (OuterVolumeSpecName: "scripts") pod "77f47a1f-3302-4d58-9b94-4ea756cc38ab" (UID: "77f47a1f-3302-4d58-9b94-4ea756cc38ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.848964 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77f47a1f-3302-4d58-9b94-4ea756cc38ab" (UID: "77f47a1f-3302-4d58-9b94-4ea756cc38ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.878843 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "77f47a1f-3302-4d58-9b94-4ea756cc38ab" (UID: "77f47a1f-3302-4d58-9b94-4ea756cc38ab"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.887280 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77f47a1f-3302-4d58-9b94-4ea756cc38ab-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.887322 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.887339 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f47a1f-3302-4d58-9b94-4ea756cc38ab-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.887350 4931 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.887364 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sljht\" (UniqueName: \"kubernetes.io/projected/77f47a1f-3302-4d58-9b94-4ea756cc38ab-kube-api-access-sljht\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.887376 4931 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77f47a1f-3302-4d58-9b94-4ea756cc38ab-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:16 crc kubenswrapper[4931]: I1129 00:20:16.887387 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77f47a1f-3302-4d58-9b94-4ea756cc38ab-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:17 crc kubenswrapper[4931]: I1129 00:20:17.060823 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:20:17 crc kubenswrapper[4931]: I1129 00:20:17.061118 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="95707751-9dd6-4fd3-a334-b9fafcccbe11" containerName="glance-log" containerID="cri-o://a8fda88fd35b023ddcfa5bdd32d75ccc4abad9b5fb0dffa414b56d66ba10dd4d" gracePeriod=30 Nov 29 00:20:17 crc kubenswrapper[4931]: I1129 00:20:17.061618 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="95707751-9dd6-4fd3-a334-b9fafcccbe11" containerName="glance-httpd" containerID="cri-o://4d5dc2f4eab8c21d8b7a78f5f16f4a198c7245fd51df7958bff094f5455d0450" gracePeriod=30 Nov 29 00:20:17 crc kubenswrapper[4931]: I1129 00:20:17.624868 4931 generic.go:334] "Generic (PLEG): container finished" podID="95707751-9dd6-4fd3-a334-b9fafcccbe11" containerID="a8fda88fd35b023ddcfa5bdd32d75ccc4abad9b5fb0dffa414b56d66ba10dd4d" exitCode=143 Nov 29 00:20:17 crc kubenswrapper[4931]: I1129 00:20:17.624925 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95707751-9dd6-4fd3-a334-b9fafcccbe11","Type":"ContainerDied","Data":"a8fda88fd35b023ddcfa5bdd32d75ccc4abad9b5fb0dffa414b56d66ba10dd4d"} Nov 29 00:20:17 crc kubenswrapper[4931]: I1129 00:20:17.627702 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-765c4d856b-6tkcn" event={"ID":"77f47a1f-3302-4d58-9b94-4ea756cc38ab","Type":"ContainerDied","Data":"87a397bd41a2a0b89231a8081f0bf70144548da1077d3da9e20a53b68b05d55d"} Nov 29 00:20:17 crc kubenswrapper[4931]: I1129 00:20:17.627743 4931 scope.go:117] "RemoveContainer" containerID="44ea44f6e434cd440fd70136642061dd183226c05674df61c397df90f3874531" Nov 29 00:20:17 crc kubenswrapper[4931]: I1129 00:20:17.627781 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-765c4d856b-6tkcn" Nov 29 00:20:17 crc kubenswrapper[4931]: I1129 00:20:17.665020 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-765c4d856b-6tkcn"] Nov 29 00:20:17 crc kubenswrapper[4931]: I1129 00:20:17.677921 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-765c4d856b-6tkcn"] Nov 29 00:20:17 crc kubenswrapper[4931]: I1129 00:20:17.806155 4931 scope.go:117] "RemoveContainer" containerID="da9b24f68267e8a0f02647614f89605bc826dc9c40f4d64d967492a86d47c0eb" Nov 29 00:20:18 crc kubenswrapper[4931]: I1129 00:20:18.279022 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:20:18 crc kubenswrapper[4931]: I1129 00:20:18.279732 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" containerName="glance-log" containerID="cri-o://3ab41f0a353db4ad8930f11915ecf91f2f4d053f82ad4c8a9cde1a3e0955ed6f" gracePeriod=30 Nov 29 00:20:18 crc kubenswrapper[4931]: I1129 00:20:18.280063 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" containerName="glance-httpd" containerID="cri-o://33a5411b3d35694169a92a5ac4550140a6c43cf93ab7ceeba00c5f40d6ca9d36" gracePeriod=30 Nov 29 00:20:18 crc kubenswrapper[4931]: I1129 00:20:18.637815 4931 generic.go:334] "Generic (PLEG): container finished" podID="ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" containerID="3ab41f0a353db4ad8930f11915ecf91f2f4d053f82ad4c8a9cde1a3e0955ed6f" exitCode=143 Nov 29 00:20:18 crc kubenswrapper[4931]: I1129 00:20:18.637847 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1","Type":"ContainerDied","Data":"3ab41f0a353db4ad8930f11915ecf91f2f4d053f82ad4c8a9cde1a3e0955ed6f"} Nov 29 00:20:18 crc kubenswrapper[4931]: I1129 00:20:18.639846 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"156abbdb-244c-407b-94b1-80c8353222f5","Type":"ContainerStarted","Data":"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f"} Nov 29 00:20:19 crc kubenswrapper[4931]: I1129 00:20:19.221223 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" path="/var/lib/kubelet/pods/77f47a1f-3302-4d58-9b94-4ea756cc38ab/volumes" Nov 29 00:20:19 crc kubenswrapper[4931]: I1129 00:20:19.946354 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:19 crc kubenswrapper[4931]: E1129 00:20:19.946494 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:20:19 crc kubenswrapper[4931]: E1129 00:20:19.947415 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-764cb44577-7q67f: configmap "swift-ring-files" not found Nov 29 00:20:19 crc kubenswrapper[4931]: E1129 00:20:19.947513 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift podName:550e16f7-f101-4a91-8e2f-45af3430d6b9 nodeName:}" failed. No retries permitted until 2025-11-29 00:20:35.947496525 +0000 UTC m=+1249.109389757 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift") pod "swift-proxy-764cb44577-7q67f" (UID: "550e16f7-f101-4a91-8e2f-45af3430d6b9") : configmap "swift-ring-files" not found Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.669556 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"156abbdb-244c-407b-94b1-80c8353222f5","Type":"ContainerStarted","Data":"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a"} Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.669934 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.669706 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="proxy-httpd" containerID="cri-o://a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a" gracePeriod=30 Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.669666 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="ceilometer-central-agent" containerID="cri-o://cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31" gracePeriod=30 Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.669756 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="sg-core" containerID="cri-o://48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f" gracePeriod=30 Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.669779 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="ceilometer-notification-agent" containerID="cri-o://8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26" gracePeriod=30 Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.672765 4931 generic.go:334] "Generic (PLEG): container finished" podID="95707751-9dd6-4fd3-a334-b9fafcccbe11" containerID="4d5dc2f4eab8c21d8b7a78f5f16f4a198c7245fd51df7958bff094f5455d0450" exitCode=0 Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.672830 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95707751-9dd6-4fd3-a334-b9fafcccbe11","Type":"ContainerDied","Data":"4d5dc2f4eab8c21d8b7a78f5f16f4a198c7245fd51df7958bff094f5455d0450"} Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.672891 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95707751-9dd6-4fd3-a334-b9fafcccbe11","Type":"ContainerDied","Data":"46fed1a5e9849dc21241f2a62f29996cf8bcb8d5d11ada3effd0764846a9ceb8"} Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.672910 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46fed1a5e9849dc21241f2a62f29996cf8bcb8d5d11ada3effd0764846a9ceb8" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.705985 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.690272369 podStartE2EDuration="13.705965172s" podCreationTimestamp="2025-11-29 00:20:07 +0000 UTC" firstStartedPulling="2025-11-29 00:20:12.687586198 +0000 UTC m=+1225.849479430" lastFinishedPulling="2025-11-29 00:20:19.703279001 +0000 UTC m=+1232.865172233" observedRunningTime="2025-11-29 00:20:20.697224059 +0000 UTC m=+1233.859117311" watchObservedRunningTime="2025-11-29 00:20:20.705965172 +0000 UTC m=+1233.867858404" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.710963 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.762784 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-scripts\") pod \"95707751-9dd6-4fd3-a334-b9fafcccbe11\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.762894 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95707751-9dd6-4fd3-a334-b9fafcccbe11-logs\") pod \"95707751-9dd6-4fd3-a334-b9fafcccbe11\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.762946 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-public-tls-certs\") pod \"95707751-9dd6-4fd3-a334-b9fafcccbe11\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.763001 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95707751-9dd6-4fd3-a334-b9fafcccbe11-httpd-run\") pod \"95707751-9dd6-4fd3-a334-b9fafcccbe11\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.763028 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkgxz\" (UniqueName: \"kubernetes.io/projected/95707751-9dd6-4fd3-a334-b9fafcccbe11-kube-api-access-bkgxz\") pod \"95707751-9dd6-4fd3-a334-b9fafcccbe11\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.763127 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-combined-ca-bundle\") pod \"95707751-9dd6-4fd3-a334-b9fafcccbe11\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.763153 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"95707751-9dd6-4fd3-a334-b9fafcccbe11\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.763172 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-config-data\") pod \"95707751-9dd6-4fd3-a334-b9fafcccbe11\" (UID: \"95707751-9dd6-4fd3-a334-b9fafcccbe11\") " Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.763769 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95707751-9dd6-4fd3-a334-b9fafcccbe11-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "95707751-9dd6-4fd3-a334-b9fafcccbe11" (UID: "95707751-9dd6-4fd3-a334-b9fafcccbe11"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.763966 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95707751-9dd6-4fd3-a334-b9fafcccbe11-logs" (OuterVolumeSpecName: "logs") pod "95707751-9dd6-4fd3-a334-b9fafcccbe11" (UID: "95707751-9dd6-4fd3-a334-b9fafcccbe11"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.769424 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "95707751-9dd6-4fd3-a334-b9fafcccbe11" (UID: "95707751-9dd6-4fd3-a334-b9fafcccbe11"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.770696 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95707751-9dd6-4fd3-a334-b9fafcccbe11-kube-api-access-bkgxz" (OuterVolumeSpecName: "kube-api-access-bkgxz") pod "95707751-9dd6-4fd3-a334-b9fafcccbe11" (UID: "95707751-9dd6-4fd3-a334-b9fafcccbe11"). InnerVolumeSpecName "kube-api-access-bkgxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.770742 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-scripts" (OuterVolumeSpecName: "scripts") pod "95707751-9dd6-4fd3-a334-b9fafcccbe11" (UID: "95707751-9dd6-4fd3-a334-b9fafcccbe11"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.821764 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95707751-9dd6-4fd3-a334-b9fafcccbe11" (UID: "95707751-9dd6-4fd3-a334-b9fafcccbe11"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.845988 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-config-data" (OuterVolumeSpecName: "config-data") pod "95707751-9dd6-4fd3-a334-b9fafcccbe11" (UID: "95707751-9dd6-4fd3-a334-b9fafcccbe11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.865760 4931 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95707751-9dd6-4fd3-a334-b9fafcccbe11-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.865798 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkgxz\" (UniqueName: \"kubernetes.io/projected/95707751-9dd6-4fd3-a334-b9fafcccbe11-kube-api-access-bkgxz\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.865811 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.865842 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.865852 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.865860 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.865868 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95707751-9dd6-4fd3-a334-b9fafcccbe11-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.884546 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.896087 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "95707751-9dd6-4fd3-a334-b9fafcccbe11" (UID: "95707751-9dd6-4fd3-a334-b9fafcccbe11"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.967578 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:20 crc kubenswrapper[4931]: I1129 00:20:20.967610 4931 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95707751-9dd6-4fd3-a334-b9fafcccbe11-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.316058 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.379687 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvr58\" (UniqueName: \"kubernetes.io/projected/156abbdb-244c-407b-94b1-80c8353222f5-kube-api-access-pvr58\") pod \"156abbdb-244c-407b-94b1-80c8353222f5\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.379752 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-scripts\") pod \"156abbdb-244c-407b-94b1-80c8353222f5\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.379940 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-sg-core-conf-yaml\") pod \"156abbdb-244c-407b-94b1-80c8353222f5\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.379975 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-config-data\") pod \"156abbdb-244c-407b-94b1-80c8353222f5\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.380049 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/156abbdb-244c-407b-94b1-80c8353222f5-log-httpd\") pod \"156abbdb-244c-407b-94b1-80c8353222f5\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.380108 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/156abbdb-244c-407b-94b1-80c8353222f5-run-httpd\") pod \"156abbdb-244c-407b-94b1-80c8353222f5\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.380133 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-combined-ca-bundle\") pod \"156abbdb-244c-407b-94b1-80c8353222f5\" (UID: \"156abbdb-244c-407b-94b1-80c8353222f5\") " Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.381047 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156abbdb-244c-407b-94b1-80c8353222f5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "156abbdb-244c-407b-94b1-80c8353222f5" (UID: "156abbdb-244c-407b-94b1-80c8353222f5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.381316 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156abbdb-244c-407b-94b1-80c8353222f5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "156abbdb-244c-407b-94b1-80c8353222f5" (UID: "156abbdb-244c-407b-94b1-80c8353222f5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.385562 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156abbdb-244c-407b-94b1-80c8353222f5-kube-api-access-pvr58" (OuterVolumeSpecName: "kube-api-access-pvr58") pod "156abbdb-244c-407b-94b1-80c8353222f5" (UID: "156abbdb-244c-407b-94b1-80c8353222f5"). InnerVolumeSpecName "kube-api-access-pvr58". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.385573 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-scripts" (OuterVolumeSpecName: "scripts") pod "156abbdb-244c-407b-94b1-80c8353222f5" (UID: "156abbdb-244c-407b-94b1-80c8353222f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.419911 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "156abbdb-244c-407b-94b1-80c8353222f5" (UID: "156abbdb-244c-407b-94b1-80c8353222f5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.474008 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "156abbdb-244c-407b-94b1-80c8353222f5" (UID: "156abbdb-244c-407b-94b1-80c8353222f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.482512 4931 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/156abbdb-244c-407b-94b1-80c8353222f5-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.482553 4931 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/156abbdb-244c-407b-94b1-80c8353222f5-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.482567 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.482581 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvr58\" (UniqueName: \"kubernetes.io/projected/156abbdb-244c-407b-94b1-80c8353222f5-kube-api-access-pvr58\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.482591 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.482601 4931 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.483818 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-config-data" (OuterVolumeSpecName: "config-data") pod "156abbdb-244c-407b-94b1-80c8353222f5" (UID: "156abbdb-244c-407b-94b1-80c8353222f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.584336 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156abbdb-244c-407b-94b1-80c8353222f5-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.685733 4931 generic.go:334] "Generic (PLEG): container finished" podID="ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" containerID="33a5411b3d35694169a92a5ac4550140a6c43cf93ab7ceeba00c5f40d6ca9d36" exitCode=0 Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.685851 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1","Type":"ContainerDied","Data":"33a5411b3d35694169a92a5ac4550140a6c43cf93ab7ceeba00c5f40d6ca9d36"} Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.697740 4931 generic.go:334] "Generic (PLEG): container finished" podID="156abbdb-244c-407b-94b1-80c8353222f5" containerID="a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a" exitCode=0 Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.697780 4931 generic.go:334] "Generic (PLEG): container finished" podID="156abbdb-244c-407b-94b1-80c8353222f5" containerID="48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f" exitCode=2 Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.697790 4931 generic.go:334] "Generic (PLEG): container finished" podID="156abbdb-244c-407b-94b1-80c8353222f5" containerID="8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26" exitCode=0 Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.697799 4931 generic.go:334] "Generic (PLEG): container finished" podID="156abbdb-244c-407b-94b1-80c8353222f5" containerID="cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31" exitCode=0 Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.697898 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.698770 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.699897 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"156abbdb-244c-407b-94b1-80c8353222f5","Type":"ContainerDied","Data":"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a"} Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.699960 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"156abbdb-244c-407b-94b1-80c8353222f5","Type":"ContainerDied","Data":"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f"} Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.699979 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"156abbdb-244c-407b-94b1-80c8353222f5","Type":"ContainerDied","Data":"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26"} Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.699991 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"156abbdb-244c-407b-94b1-80c8353222f5","Type":"ContainerDied","Data":"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31"} Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.700005 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"156abbdb-244c-407b-94b1-80c8353222f5","Type":"ContainerDied","Data":"4fbc00259382e453e90415fcc7e096975e8abcd9ffc553bdcb3947360b85da6c"} Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.700026 4931 scope.go:117] "RemoveContainer" containerID="a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.741296 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.742877 4931 scope.go:117] "RemoveContainer" containerID="48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.752489 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.800737 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:20:21 crc kubenswrapper[4931]: E1129 00:20:21.801405 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="proxy-httpd" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.801430 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="proxy-httpd" Nov 29 00:20:21 crc kubenswrapper[4931]: E1129 00:20:21.801446 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="ceilometer-central-agent" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.801454 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="ceilometer-central-agent" Nov 29 00:20:21 crc kubenswrapper[4931]: E1129 00:20:21.801463 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon-log" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.801472 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon-log" Nov 29 00:20:21 crc kubenswrapper[4931]: E1129 00:20:21.801484 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="sg-core" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.801492 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="sg-core" Nov 29 00:20:21 crc kubenswrapper[4931]: E1129 00:20:21.801509 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="ceilometer-notification-agent" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.801517 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="ceilometer-notification-agent" Nov 29 00:20:21 crc kubenswrapper[4931]: E1129 00:20:21.801532 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95707751-9dd6-4fd3-a334-b9fafcccbe11" containerName="glance-log" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.801539 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="95707751-9dd6-4fd3-a334-b9fafcccbe11" containerName="glance-log" Nov 29 00:20:21 crc kubenswrapper[4931]: E1129 00:20:21.801564 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.801574 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon" Nov 29 00:20:21 crc kubenswrapper[4931]: E1129 00:20:21.801696 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95707751-9dd6-4fd3-a334-b9fafcccbe11" containerName="glance-httpd" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.801708 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="95707751-9dd6-4fd3-a334-b9fafcccbe11" containerName="glance-httpd" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.802089 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="ceilometer-central-agent" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.802113 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="95707751-9dd6-4fd3-a334-b9fafcccbe11" containerName="glance-httpd" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.802128 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="ceilometer-notification-agent" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.802151 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="sg-core" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.802169 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon-log" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.802184 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="156abbdb-244c-407b-94b1-80c8353222f5" containerName="proxy-httpd" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.802196 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="95707751-9dd6-4fd3-a334-b9fafcccbe11" containerName="glance-log" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.802207 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f47a1f-3302-4d58-9b94-4ea756cc38ab" containerName="horizon" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.803909 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.810582 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.811151 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.825423 4931 scope.go:117] "RemoveContainer" containerID="8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.829348 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.866891 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.882923 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.890174 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.890258 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.890297 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-config-data\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.890318 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-scripts\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.890333 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-logs\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.890395 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.890416 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzrxp\" (UniqueName: \"kubernetes.io/projected/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-kube-api-access-qzrxp\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.890437 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.910078 4931 scope.go:117] "RemoveContainer" containerID="cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.912243 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.914751 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.917516 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.917576 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.925097 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.992676 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.992742 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.992767 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-config-data\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.992892 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.992913 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-config-data\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.992935 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-scripts\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.992967 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-logs\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.992989 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79df5ada-b426-413e-a995-287294286bed-log-httpd\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.993035 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.993059 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69clr\" (UniqueName: \"kubernetes.io/projected/79df5ada-b426-413e-a995-287294286bed-kube-api-access-69clr\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.993086 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.993105 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzrxp\" (UniqueName: \"kubernetes.io/projected/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-kube-api-access-qzrxp\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.993126 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.993165 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79df5ada-b426-413e-a995-287294286bed-run-httpd\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.993189 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-scripts\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.993621 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.995375 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:21 crc kubenswrapper[4931]: I1129 00:20:21.996858 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-logs\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.000390 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-scripts\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.003623 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-config-data\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.005029 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.009123 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.009522 4931 scope.go:117] "RemoveContainer" containerID="a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a" Nov 29 00:20:22 crc kubenswrapper[4931]: E1129 00:20:22.009935 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a\": container with ID starting with a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a not found: ID does not exist" containerID="a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.009963 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a"} err="failed to get container status \"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a\": rpc error: code = NotFound desc = could not find container \"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a\": container with ID starting with a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.009985 4931 scope.go:117] "RemoveContainer" containerID="48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f" Nov 29 00:20:22 crc kubenswrapper[4931]: E1129 00:20:22.010217 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f\": container with ID starting with 48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f not found: ID does not exist" containerID="48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.010244 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f"} err="failed to get container status \"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f\": rpc error: code = NotFound desc = could not find container \"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f\": container with ID starting with 48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.010260 4931 scope.go:117] "RemoveContainer" containerID="8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26" Nov 29 00:20:22 crc kubenswrapper[4931]: E1129 00:20:22.010630 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26\": container with ID starting with 8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26 not found: ID does not exist" containerID="8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.010674 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26"} err="failed to get container status \"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26\": rpc error: code = NotFound desc = could not find container \"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26\": container with ID starting with 8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26 not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.010696 4931 scope.go:117] "RemoveContainer" containerID="cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31" Nov 29 00:20:22 crc kubenswrapper[4931]: E1129 00:20:22.011035 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31\": container with ID starting with cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31 not found: ID does not exist" containerID="cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.011068 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31"} err="failed to get container status \"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31\": rpc error: code = NotFound desc = could not find container \"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31\": container with ID starting with cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31 not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.011086 4931 scope.go:117] "RemoveContainer" containerID="a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.011413 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a"} err="failed to get container status \"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a\": rpc error: code = NotFound desc = could not find container \"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a\": container with ID starting with a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.011464 4931 scope.go:117] "RemoveContainer" containerID="48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.014963 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f"} err="failed to get container status \"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f\": rpc error: code = NotFound desc = could not find container \"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f\": container with ID starting with 48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.015001 4931 scope.go:117] "RemoveContainer" containerID="8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.015340 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26"} err="failed to get container status \"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26\": rpc error: code = NotFound desc = could not find container \"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26\": container with ID starting with 8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26 not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.015393 4931 scope.go:117] "RemoveContainer" containerID="cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.015674 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31"} err="failed to get container status \"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31\": rpc error: code = NotFound desc = could not find container \"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31\": container with ID starting with cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31 not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.015717 4931 scope.go:117] "RemoveContainer" containerID="a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.015967 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzrxp\" (UniqueName: \"kubernetes.io/projected/ea6acd51-8e93-4fdd-9535-84c7e6d2cee6-kube-api-access-qzrxp\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.016046 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a"} err="failed to get container status \"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a\": rpc error: code = NotFound desc = could not find container \"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a\": container with ID starting with a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.016075 4931 scope.go:117] "RemoveContainer" containerID="48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.018093 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f"} err="failed to get container status \"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f\": rpc error: code = NotFound desc = could not find container \"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f\": container with ID starting with 48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.018137 4931 scope.go:117] "RemoveContainer" containerID="8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.018461 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26"} err="failed to get container status \"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26\": rpc error: code = NotFound desc = could not find container \"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26\": container with ID starting with 8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26 not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.018493 4931 scope.go:117] "RemoveContainer" containerID="cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.018869 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31"} err="failed to get container status \"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31\": rpc error: code = NotFound desc = could not find container \"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31\": container with ID starting with cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31 not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.018905 4931 scope.go:117] "RemoveContainer" containerID="a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.019935 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a"} err="failed to get container status \"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a\": rpc error: code = NotFound desc = could not find container \"a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a\": container with ID starting with a6437ab5a8ff5024ca13c1caece6df7efd6b7b509d94d997b834865d1357f06a not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.019975 4931 scope.go:117] "RemoveContainer" containerID="48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.020255 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f"} err="failed to get container status \"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f\": rpc error: code = NotFound desc = could not find container \"48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f\": container with ID starting with 48d333e7d593472ca5dfae4b33600385cff3c3d92d6076d530f3b64d01acb01f not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.020282 4931 scope.go:117] "RemoveContainer" containerID="8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.020588 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26"} err="failed to get container status \"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26\": rpc error: code = NotFound desc = could not find container \"8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26\": container with ID starting with 8f7e0f615ec23a712142ae82a95d8a2a1359d98fd9566fa685c3718abc814e26 not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.020628 4931 scope.go:117] "RemoveContainer" containerID="cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.022276 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31"} err="failed to get container status \"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31\": rpc error: code = NotFound desc = could not find container \"cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31\": container with ID starting with cb389b5a6c125b346eb3e19fc353f83647c43a7850a09af481fe87ec1700ba31 not found: ID does not exist" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.031426 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6\") " pod="openstack/glance-default-external-api-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.087710 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.095095 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-config-data\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.095141 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.095187 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79df5ada-b426-413e-a995-287294286bed-log-httpd\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.095264 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.095299 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69clr\" (UniqueName: \"kubernetes.io/projected/79df5ada-b426-413e-a995-287294286bed-kube-api-access-69clr\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.095365 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79df5ada-b426-413e-a995-287294286bed-run-httpd\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.095399 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-scripts\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.096357 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79df5ada-b426-413e-a995-287294286bed-log-httpd\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.096943 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79df5ada-b426-413e-a995-287294286bed-run-httpd\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.099852 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-config-data\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.100507 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.103023 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-scripts\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.103162 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.123312 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69clr\" (UniqueName: \"kubernetes.io/projected/79df5ada-b426-413e-a995-287294286bed-kube-api-access-69clr\") pod \"ceilometer-0\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.147259 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.198201 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-config-data\") pod \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.198274 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-httpd-run\") pod \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.198331 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-logs\") pod \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.198367 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfqg5\" (UniqueName: \"kubernetes.io/projected/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-kube-api-access-mfqg5\") pod \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.198425 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.198466 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-combined-ca-bundle\") pod \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.198539 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-scripts\") pod \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.198662 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-internal-tls-certs\") pod \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\" (UID: \"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1\") " Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.200736 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" (UID: "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.202066 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-logs" (OuterVolumeSpecName: "logs") pod "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" (UID: "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.203992 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" (UID: "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.207981 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-kube-api-access-mfqg5" (OuterVolumeSpecName: "kube-api-access-mfqg5") pod "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" (UID: "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1"). InnerVolumeSpecName "kube-api-access-mfqg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.211280 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-scripts" (OuterVolumeSpecName: "scripts") pod "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" (UID: "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.232774 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" (UID: "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.286012 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-config-data" (OuterVolumeSpecName: "config-data") pod "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" (UID: "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.307116 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.307151 4931 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.307163 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.307175 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfqg5\" (UniqueName: \"kubernetes.io/projected/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-kube-api-access-mfqg5\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.307200 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.307213 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.307225 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.315910 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" (UID: "ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.335114 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.387311 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.408928 4931 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.408954 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.717538 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1","Type":"ContainerDied","Data":"c27a49a679d4578f60ec9719f2f389e5bb4cfa24fb3a156b4e174b91a8ff51d0"} Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.717629 4931 scope.go:117] "RemoveContainer" containerID="33a5411b3d35694169a92a5ac4550140a6c43cf93ab7ceeba00c5f40d6ca9d36" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.717587 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:22 crc kubenswrapper[4931]: W1129 00:20:22.725928 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea6acd51_8e93_4fdd_9535_84c7e6d2cee6.slice/crio-7c6ff709f2fd219600de6864fa6a90899030c170572d77c6c7b93b7d887b9f6a WatchSource:0}: Error finding container 7c6ff709f2fd219600de6864fa6a90899030c170572d77c6c7b93b7d887b9f6a: Status 404 returned error can't find the container with id 7c6ff709f2fd219600de6864fa6a90899030c170572d77c6c7b93b7d887b9f6a Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.728254 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.840044 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.862729 4931 scope.go:117] "RemoveContainer" containerID="3ab41f0a353db4ad8930f11915ecf91f2f4d053f82ad4c8a9cde1a3e0955ed6f" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.867922 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.887879 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.915264 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:20:22 crc kubenswrapper[4931]: E1129 00:20:22.915803 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" containerName="glance-httpd" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.915839 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" containerName="glance-httpd" Nov 29 00:20:22 crc kubenswrapper[4931]: E1129 00:20:22.915869 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" containerName="glance-log" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.915878 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" containerName="glance-log" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.916105 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" containerName="glance-httpd" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.916140 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" containerName="glance-log" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.918044 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.921214 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.921302 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 29 00:20:22 crc kubenswrapper[4931]: I1129 00:20:22.934677 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.020019 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78acdbbf-4fab-4641-86e8-49f124b8247e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.020187 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/78acdbbf-4fab-4641-86e8-49f124b8247e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.020349 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78acdbbf-4fab-4641-86e8-49f124b8247e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.020416 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78acdbbf-4fab-4641-86e8-49f124b8247e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.020441 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.020459 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78acdbbf-4fab-4641-86e8-49f124b8247e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.020489 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78acdbbf-4fab-4641-86e8-49f124b8247e-logs\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.020744 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb2v8\" (UniqueName: \"kubernetes.io/projected/78acdbbf-4fab-4641-86e8-49f124b8247e-kube-api-access-jb2v8\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.122078 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78acdbbf-4fab-4641-86e8-49f124b8247e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.122144 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/78acdbbf-4fab-4641-86e8-49f124b8247e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.122203 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78acdbbf-4fab-4641-86e8-49f124b8247e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.122243 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78acdbbf-4fab-4641-86e8-49f124b8247e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.122272 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.122295 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78acdbbf-4fab-4641-86e8-49f124b8247e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.122320 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78acdbbf-4fab-4641-86e8-49f124b8247e-logs\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.122360 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb2v8\" (UniqueName: \"kubernetes.io/projected/78acdbbf-4fab-4641-86e8-49f124b8247e-kube-api-access-jb2v8\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.122813 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/78acdbbf-4fab-4641-86e8-49f124b8247e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.122894 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.123167 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78acdbbf-4fab-4641-86e8-49f124b8247e-logs\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.129254 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78acdbbf-4fab-4641-86e8-49f124b8247e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.129289 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78acdbbf-4fab-4641-86e8-49f124b8247e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.130552 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78acdbbf-4fab-4641-86e8-49f124b8247e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.138599 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78acdbbf-4fab-4641-86e8-49f124b8247e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.144810 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb2v8\" (UniqueName: \"kubernetes.io/projected/78acdbbf-4fab-4641-86e8-49f124b8247e-kube-api-access-jb2v8\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.156586 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"78acdbbf-4fab-4641-86e8-49f124b8247e\") " pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.222474 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="156abbdb-244c-407b-94b1-80c8353222f5" path="/var/lib/kubelet/pods/156abbdb-244c-407b-94b1-80c8353222f5/volumes" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.223406 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95707751-9dd6-4fd3-a334-b9fafcccbe11" path="/var/lib/kubelet/pods/95707751-9dd6-4fd3-a334-b9fafcccbe11/volumes" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.224538 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1" path="/var/lib/kubelet/pods/ddfe2eed-f8f8-43c7-99ef-92306bd9c1e1/volumes" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.253532 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.735565 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.736147 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.742330 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6","Type":"ContainerStarted","Data":"3519c9f91f948ca0d952f574e7d05cef41d6754b82cce9c36a96fe9c2742c983"} Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.742389 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6","Type":"ContainerStarted","Data":"7c6ff709f2fd219600de6864fa6a90899030c170572d77c6c7b93b7d887b9f6a"} Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.751405 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79df5ada-b426-413e-a995-287294286bed","Type":"ContainerStarted","Data":"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c"} Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.751446 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79df5ada-b426-413e-a995-287294286bed","Type":"ContainerStarted","Data":"962a04bd5eb9996f20e493391810e1cc5721a395aa04ffaba05d6ac21b8f1dfa"} Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.798974 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 29 00:20:23 crc kubenswrapper[4931]: W1129 00:20:23.801636 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78acdbbf_4fab_4641_86e8_49f124b8247e.slice/crio-ec0cac14bf741df950796d1bfe759d85c67951dd7245a9197e9f282d4f1c87f0 WatchSource:0}: Error finding container ec0cac14bf741df950796d1bfe759d85c67951dd7245a9197e9f282d4f1c87f0: Status 404 returned error can't find the container with id ec0cac14bf741df950796d1bfe759d85c67951dd7245a9197e9f282d4f1c87f0 Nov 29 00:20:23 crc kubenswrapper[4931]: I1129 00:20:23.843256 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:24 crc kubenswrapper[4931]: I1129 00:20:24.765974 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79df5ada-b426-413e-a995-287294286bed","Type":"ContainerStarted","Data":"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9"} Nov 29 00:20:24 crc kubenswrapper[4931]: I1129 00:20:24.768995 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"78acdbbf-4fab-4641-86e8-49f124b8247e","Type":"ContainerStarted","Data":"9039a7ac09e754f1b90f7ffa88c055e3a37ec5acb6e8a9635e1d4c7e8d2032d2"} Nov 29 00:20:24 crc kubenswrapper[4931]: I1129 00:20:24.769035 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"78acdbbf-4fab-4641-86e8-49f124b8247e","Type":"ContainerStarted","Data":"ec0cac14bf741df950796d1bfe759d85c67951dd7245a9197e9f282d4f1c87f0"} Nov 29 00:20:24 crc kubenswrapper[4931]: I1129 00:20:24.772065 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ea6acd51-8e93-4fdd-9535-84c7e6d2cee6","Type":"ContainerStarted","Data":"ed989d45fa85a889c51d3a104314f16c1818ae0c3e8e78467e26cdddfba514b0"} Nov 29 00:20:24 crc kubenswrapper[4931]: I1129 00:20:24.793534 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.793515577 podStartE2EDuration="3.793515577s" podCreationTimestamp="2025-11-29 00:20:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:20:24.793130346 +0000 UTC m=+1237.955023598" watchObservedRunningTime="2025-11-29 00:20:24.793515577 +0000 UTC m=+1237.955408809" Nov 29 00:20:25 crc kubenswrapper[4931]: I1129 00:20:25.781329 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"78acdbbf-4fab-4641-86e8-49f124b8247e","Type":"ContainerStarted","Data":"5af1de70696292f435a5221b0b0cc5c555cb69fbb9db8f72a3349e121a4d631e"} Nov 29 00:20:25 crc kubenswrapper[4931]: I1129 00:20:25.787033 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79df5ada-b426-413e-a995-287294286bed","Type":"ContainerStarted","Data":"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3"} Nov 29 00:20:25 crc kubenswrapper[4931]: I1129 00:20:25.805173 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.805156347 podStartE2EDuration="3.805156347s" podCreationTimestamp="2025-11-29 00:20:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:20:25.802360316 +0000 UTC m=+1238.964253558" watchObservedRunningTime="2025-11-29 00:20:25.805156347 +0000 UTC m=+1238.967049579" Nov 29 00:20:27 crc kubenswrapper[4931]: I1129 00:20:27.809195 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79df5ada-b426-413e-a995-287294286bed","Type":"ContainerStarted","Data":"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3"} Nov 29 00:20:27 crc kubenswrapper[4931]: I1129 00:20:27.809454 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="ceilometer-central-agent" containerID="cri-o://685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c" gracePeriod=30 Nov 29 00:20:27 crc kubenswrapper[4931]: I1129 00:20:27.809916 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="proxy-httpd" containerID="cri-o://89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3" gracePeriod=30 Nov 29 00:20:27 crc kubenswrapper[4931]: I1129 00:20:27.810007 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="ceilometer-notification-agent" containerID="cri-o://d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9" gracePeriod=30 Nov 29 00:20:27 crc kubenswrapper[4931]: I1129 00:20:27.810053 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="sg-core" containerID="cri-o://9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3" gracePeriod=30 Nov 29 00:20:27 crc kubenswrapper[4931]: I1129 00:20:27.810132 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 00:20:27 crc kubenswrapper[4931]: I1129 00:20:27.835266 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.780682304 podStartE2EDuration="6.835242964s" podCreationTimestamp="2025-11-29 00:20:21 +0000 UTC" firstStartedPulling="2025-11-29 00:20:22.878895765 +0000 UTC m=+1236.040788997" lastFinishedPulling="2025-11-29 00:20:26.933456425 +0000 UTC m=+1240.095349657" observedRunningTime="2025-11-29 00:20:27.826519201 +0000 UTC m=+1240.988412453" watchObservedRunningTime="2025-11-29 00:20:27.835242964 +0000 UTC m=+1240.997136196" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.061879 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-r7n7s"] Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.064335 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-r7n7s" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.091377 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-r7n7s"] Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.129496 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-n899b"] Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.132221 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n899b" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.147085 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-n899b"] Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.242802 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc9l9\" (UniqueName: \"kubernetes.io/projected/6284f70c-aaa2-4182-98fd-82fb59295e1f-kube-api-access-mc9l9\") pod \"nova-cell0-db-create-n899b\" (UID: \"6284f70c-aaa2-4182-98fd-82fb59295e1f\") " pod="openstack/nova-cell0-db-create-n899b" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.243106 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t5hv\" (UniqueName: \"kubernetes.io/projected/2f6ca58a-4bc4-4277-891d-db214595cb56-kube-api-access-7t5hv\") pod \"nova-api-db-create-r7n7s\" (UID: \"2f6ca58a-4bc4-4277-891d-db214595cb56\") " pod="openstack/nova-api-db-create-r7n7s" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.243157 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6284f70c-aaa2-4182-98fd-82fb59295e1f-operator-scripts\") pod \"nova-cell0-db-create-n899b\" (UID: \"6284f70c-aaa2-4182-98fd-82fb59295e1f\") " pod="openstack/nova-cell0-db-create-n899b" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.243209 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f6ca58a-4bc4-4277-891d-db214595cb56-operator-scripts\") pod \"nova-api-db-create-r7n7s\" (UID: \"2f6ca58a-4bc4-4277-891d-db214595cb56\") " pod="openstack/nova-api-db-create-r7n7s" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.243301 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-68b1-account-create-update-4flsl"] Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.244363 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-68b1-account-create-update-4flsl" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.247615 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.251288 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-68b1-account-create-update-4flsl"] Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.344984 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t5hv\" (UniqueName: \"kubernetes.io/projected/2f6ca58a-4bc4-4277-891d-db214595cb56-kube-api-access-7t5hv\") pod \"nova-api-db-create-r7n7s\" (UID: \"2f6ca58a-4bc4-4277-891d-db214595cb56\") " pod="openstack/nova-api-db-create-r7n7s" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.345060 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6284f70c-aaa2-4182-98fd-82fb59295e1f-operator-scripts\") pod \"nova-cell0-db-create-n899b\" (UID: \"6284f70c-aaa2-4182-98fd-82fb59295e1f\") " pod="openstack/nova-cell0-db-create-n899b" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.345117 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f6ca58a-4bc4-4277-891d-db214595cb56-operator-scripts\") pod \"nova-api-db-create-r7n7s\" (UID: \"2f6ca58a-4bc4-4277-891d-db214595cb56\") " pod="openstack/nova-api-db-create-r7n7s" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.345156 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7tqx\" (UniqueName: \"kubernetes.io/projected/46687ab2-074a-4007-b89a-a962c9b1d14d-kube-api-access-h7tqx\") pod \"nova-api-68b1-account-create-update-4flsl\" (UID: \"46687ab2-074a-4007-b89a-a962c9b1d14d\") " pod="openstack/nova-api-68b1-account-create-update-4flsl" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.345193 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46687ab2-074a-4007-b89a-a962c9b1d14d-operator-scripts\") pod \"nova-api-68b1-account-create-update-4flsl\" (UID: \"46687ab2-074a-4007-b89a-a962c9b1d14d\") " pod="openstack/nova-api-68b1-account-create-update-4flsl" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.345211 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc9l9\" (UniqueName: \"kubernetes.io/projected/6284f70c-aaa2-4182-98fd-82fb59295e1f-kube-api-access-mc9l9\") pod \"nova-cell0-db-create-n899b\" (UID: \"6284f70c-aaa2-4182-98fd-82fb59295e1f\") " pod="openstack/nova-cell0-db-create-n899b" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.346081 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f6ca58a-4bc4-4277-891d-db214595cb56-operator-scripts\") pod \"nova-api-db-create-r7n7s\" (UID: \"2f6ca58a-4bc4-4277-891d-db214595cb56\") " pod="openstack/nova-api-db-create-r7n7s" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.354924 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6284f70c-aaa2-4182-98fd-82fb59295e1f-operator-scripts\") pod \"nova-cell0-db-create-n899b\" (UID: \"6284f70c-aaa2-4182-98fd-82fb59295e1f\") " pod="openstack/nova-cell0-db-create-n899b" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.360468 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-dw4x9"] Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.361848 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dw4x9" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.365599 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc9l9\" (UniqueName: \"kubernetes.io/projected/6284f70c-aaa2-4182-98fd-82fb59295e1f-kube-api-access-mc9l9\") pod \"nova-cell0-db-create-n899b\" (UID: \"6284f70c-aaa2-4182-98fd-82fb59295e1f\") " pod="openstack/nova-cell0-db-create-n899b" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.365725 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t5hv\" (UniqueName: \"kubernetes.io/projected/2f6ca58a-4bc4-4277-891d-db214595cb56-kube-api-access-7t5hv\") pod \"nova-api-db-create-r7n7s\" (UID: \"2f6ca58a-4bc4-4277-891d-db214595cb56\") " pod="openstack/nova-api-db-create-r7n7s" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.369843 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dw4x9"] Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.392490 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-r7n7s" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.446497 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7tqx\" (UniqueName: \"kubernetes.io/projected/46687ab2-074a-4007-b89a-a962c9b1d14d-kube-api-access-h7tqx\") pod \"nova-api-68b1-account-create-update-4flsl\" (UID: \"46687ab2-074a-4007-b89a-a962c9b1d14d\") " pod="openstack/nova-api-68b1-account-create-update-4flsl" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.446561 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46687ab2-074a-4007-b89a-a962c9b1d14d-operator-scripts\") pod \"nova-api-68b1-account-create-update-4flsl\" (UID: \"46687ab2-074a-4007-b89a-a962c9b1d14d\") " pod="openstack/nova-api-68b1-account-create-update-4flsl" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.447336 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46687ab2-074a-4007-b89a-a962c9b1d14d-operator-scripts\") pod \"nova-api-68b1-account-create-update-4flsl\" (UID: \"46687ab2-074a-4007-b89a-a962c9b1d14d\") " pod="openstack/nova-api-68b1-account-create-update-4flsl" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.450329 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-3603-account-create-update-8rgkv"] Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.451466 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3603-account-create-update-8rgkv" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.455299 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.459782 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n899b" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.459997 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3603-account-create-update-8rgkv"] Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.477937 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7tqx\" (UniqueName: \"kubernetes.io/projected/46687ab2-074a-4007-b89a-a962c9b1d14d-kube-api-access-h7tqx\") pod \"nova-api-68b1-account-create-update-4flsl\" (UID: \"46687ab2-074a-4007-b89a-a962c9b1d14d\") " pod="openstack/nova-api-68b1-account-create-update-4flsl" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.551238 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5chdf\" (UniqueName: \"kubernetes.io/projected/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5-kube-api-access-5chdf\") pod \"nova-cell1-db-create-dw4x9\" (UID: \"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5\") " pod="openstack/nova-cell1-db-create-dw4x9" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.551538 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5-operator-scripts\") pod \"nova-cell1-db-create-dw4x9\" (UID: \"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5\") " pod="openstack/nova-cell1-db-create-dw4x9" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.551568 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a57e64d1-340e-4e80-8112-16625e6d7d6e-operator-scripts\") pod \"nova-cell0-3603-account-create-update-8rgkv\" (UID: \"a57e64d1-340e-4e80-8112-16625e6d7d6e\") " pod="openstack/nova-cell0-3603-account-create-update-8rgkv" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.551591 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6zdh\" (UniqueName: \"kubernetes.io/projected/a57e64d1-340e-4e80-8112-16625e6d7d6e-kube-api-access-x6zdh\") pod \"nova-cell0-3603-account-create-update-8rgkv\" (UID: \"a57e64d1-340e-4e80-8112-16625e6d7d6e\") " pod="openstack/nova-cell0-3603-account-create-update-8rgkv" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.639704 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-4b32-account-create-update-nhrbj"] Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.640928 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.643046 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.653756 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5chdf\" (UniqueName: \"kubernetes.io/projected/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5-kube-api-access-5chdf\") pod \"nova-cell1-db-create-dw4x9\" (UID: \"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5\") " pod="openstack/nova-cell1-db-create-dw4x9" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.653800 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5-operator-scripts\") pod \"nova-cell1-db-create-dw4x9\" (UID: \"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5\") " pod="openstack/nova-cell1-db-create-dw4x9" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.653843 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a57e64d1-340e-4e80-8112-16625e6d7d6e-operator-scripts\") pod \"nova-cell0-3603-account-create-update-8rgkv\" (UID: \"a57e64d1-340e-4e80-8112-16625e6d7d6e\") " pod="openstack/nova-cell0-3603-account-create-update-8rgkv" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.653869 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6zdh\" (UniqueName: \"kubernetes.io/projected/a57e64d1-340e-4e80-8112-16625e6d7d6e-kube-api-access-x6zdh\") pod \"nova-cell0-3603-account-create-update-8rgkv\" (UID: \"a57e64d1-340e-4e80-8112-16625e6d7d6e\") " pod="openstack/nova-cell0-3603-account-create-update-8rgkv" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.654749 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5-operator-scripts\") pod \"nova-cell1-db-create-dw4x9\" (UID: \"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5\") " pod="openstack/nova-cell1-db-create-dw4x9" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.655881 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a57e64d1-340e-4e80-8112-16625e6d7d6e-operator-scripts\") pod \"nova-cell0-3603-account-create-update-8rgkv\" (UID: \"a57e64d1-340e-4e80-8112-16625e6d7d6e\") " pod="openstack/nova-cell0-3603-account-create-update-8rgkv" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.658018 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4b32-account-create-update-nhrbj"] Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.671883 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6zdh\" (UniqueName: \"kubernetes.io/projected/a57e64d1-340e-4e80-8112-16625e6d7d6e-kube-api-access-x6zdh\") pod \"nova-cell0-3603-account-create-update-8rgkv\" (UID: \"a57e64d1-340e-4e80-8112-16625e6d7d6e\") " pod="openstack/nova-cell0-3603-account-create-update-8rgkv" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.672184 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5chdf\" (UniqueName: \"kubernetes.io/projected/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5-kube-api-access-5chdf\") pod \"nova-cell1-db-create-dw4x9\" (UID: \"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5\") " pod="openstack/nova-cell1-db-create-dw4x9" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.708993 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.720504 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-68b1-account-create-update-4flsl" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.732522 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dw4x9" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.755211 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f82d911-133a-4c7a-87a4-9684bb066cc0-operator-scripts\") pod \"nova-cell1-4b32-account-create-update-nhrbj\" (UID: \"5f82d911-133a-4c7a-87a4-9684bb066cc0\") " pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.755352 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g4kg\" (UniqueName: \"kubernetes.io/projected/5f82d911-133a-4c7a-87a4-9684bb066cc0-kube-api-access-5g4kg\") pod \"nova-cell1-4b32-account-create-update-nhrbj\" (UID: \"5f82d911-133a-4c7a-87a4-9684bb066cc0\") " pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.836315 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3603-account-create-update-8rgkv" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.853294 4931 generic.go:334] "Generic (PLEG): container finished" podID="79df5ada-b426-413e-a995-287294286bed" containerID="89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3" exitCode=0 Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.853330 4931 generic.go:334] "Generic (PLEG): container finished" podID="79df5ada-b426-413e-a995-287294286bed" containerID="9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3" exitCode=2 Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.853339 4931 generic.go:334] "Generic (PLEG): container finished" podID="79df5ada-b426-413e-a995-287294286bed" containerID="d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9" exitCode=0 Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.853347 4931 generic.go:334] "Generic (PLEG): container finished" podID="79df5ada-b426-413e-a995-287294286bed" containerID="685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c" exitCode=0 Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.853368 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79df5ada-b426-413e-a995-287294286bed","Type":"ContainerDied","Data":"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3"} Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.853398 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79df5ada-b426-413e-a995-287294286bed","Type":"ContainerDied","Data":"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3"} Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.853410 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79df5ada-b426-413e-a995-287294286bed","Type":"ContainerDied","Data":"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9"} Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.853422 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79df5ada-b426-413e-a995-287294286bed","Type":"ContainerDied","Data":"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c"} Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.853432 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"79df5ada-b426-413e-a995-287294286bed","Type":"ContainerDied","Data":"962a04bd5eb9996f20e493391810e1cc5721a395aa04ffaba05d6ac21b8f1dfa"} Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.853449 4931 scope.go:117] "RemoveContainer" containerID="89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.853602 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.856388 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-combined-ca-bundle\") pod \"79df5ada-b426-413e-a995-287294286bed\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.856493 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79df5ada-b426-413e-a995-287294286bed-run-httpd\") pod \"79df5ada-b426-413e-a995-287294286bed\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.856540 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79df5ada-b426-413e-a995-287294286bed-log-httpd\") pod \"79df5ada-b426-413e-a995-287294286bed\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.856565 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-sg-core-conf-yaml\") pod \"79df5ada-b426-413e-a995-287294286bed\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.856613 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69clr\" (UniqueName: \"kubernetes.io/projected/79df5ada-b426-413e-a995-287294286bed-kube-api-access-69clr\") pod \"79df5ada-b426-413e-a995-287294286bed\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.856652 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-scripts\") pod \"79df5ada-b426-413e-a995-287294286bed\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.856755 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-config-data\") pod \"79df5ada-b426-413e-a995-287294286bed\" (UID: \"79df5ada-b426-413e-a995-287294286bed\") " Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.857111 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f82d911-133a-4c7a-87a4-9684bb066cc0-operator-scripts\") pod \"nova-cell1-4b32-account-create-update-nhrbj\" (UID: \"5f82d911-133a-4c7a-87a4-9684bb066cc0\") " pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.857225 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g4kg\" (UniqueName: \"kubernetes.io/projected/5f82d911-133a-4c7a-87a4-9684bb066cc0-kube-api-access-5g4kg\") pod \"nova-cell1-4b32-account-create-update-nhrbj\" (UID: \"5f82d911-133a-4c7a-87a4-9684bb066cc0\") " pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.861664 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f82d911-133a-4c7a-87a4-9684bb066cc0-operator-scripts\") pod \"nova-cell1-4b32-account-create-update-nhrbj\" (UID: \"5f82d911-133a-4c7a-87a4-9684bb066cc0\") " pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.861932 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79df5ada-b426-413e-a995-287294286bed-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "79df5ada-b426-413e-a995-287294286bed" (UID: "79df5ada-b426-413e-a995-287294286bed"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.863776 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79df5ada-b426-413e-a995-287294286bed-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "79df5ada-b426-413e-a995-287294286bed" (UID: "79df5ada-b426-413e-a995-287294286bed"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.864134 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79df5ada-b426-413e-a995-287294286bed-kube-api-access-69clr" (OuterVolumeSpecName: "kube-api-access-69clr") pod "79df5ada-b426-413e-a995-287294286bed" (UID: "79df5ada-b426-413e-a995-287294286bed"). InnerVolumeSpecName "kube-api-access-69clr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.888962 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-scripts" (OuterVolumeSpecName: "scripts") pod "79df5ada-b426-413e-a995-287294286bed" (UID: "79df5ada-b426-413e-a995-287294286bed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.943214 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "79df5ada-b426-413e-a995-287294286bed" (UID: "79df5ada-b426-413e-a995-287294286bed"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.947607 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g4kg\" (UniqueName: \"kubernetes.io/projected/5f82d911-133a-4c7a-87a4-9684bb066cc0-kube-api-access-5g4kg\") pod \"nova-cell1-4b32-account-create-update-nhrbj\" (UID: \"5f82d911-133a-4c7a-87a4-9684bb066cc0\") " pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.961450 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.963296 4931 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79df5ada-b426-413e-a995-287294286bed-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.963330 4931 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/79df5ada-b426-413e-a995-287294286bed-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.963342 4931 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.963354 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69clr\" (UniqueName: \"kubernetes.io/projected/79df5ada-b426-413e-a995-287294286bed-kube-api-access-69clr\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.963365 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:28 crc kubenswrapper[4931]: I1129 00:20:28.989421 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-r7n7s"] Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.045130 4931 scope.go:117] "RemoveContainer" containerID="9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.060964 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79df5ada-b426-413e-a995-287294286bed" (UID: "79df5ada-b426-413e-a995-287294286bed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.064618 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.102489 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-n899b"] Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.145326 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-config-data" (OuterVolumeSpecName: "config-data") pod "79df5ada-b426-413e-a995-287294286bed" (UID: "79df5ada-b426-413e-a995-287294286bed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.166664 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79df5ada-b426-413e-a995-287294286bed-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.249670 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.259041 4931 scope.go:117] "RemoveContainer" containerID="d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.266712 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.275883 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:29 crc kubenswrapper[4931]: E1129 00:20:29.276408 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="ceilometer-central-agent" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.276427 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="ceilometer-central-agent" Nov 29 00:20:29 crc kubenswrapper[4931]: E1129 00:20:29.276444 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="sg-core" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.276450 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="sg-core" Nov 29 00:20:29 crc kubenswrapper[4931]: E1129 00:20:29.276478 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="proxy-httpd" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.276484 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="proxy-httpd" Nov 29 00:20:29 crc kubenswrapper[4931]: E1129 00:20:29.276494 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="ceilometer-notification-agent" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.276499 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="ceilometer-notification-agent" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.276662 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="proxy-httpd" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.276672 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="sg-core" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.276687 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="ceilometer-notification-agent" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.276711 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="79df5ada-b426-413e-a995-287294286bed" containerName="ceilometer-central-agent" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.290592 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.291049 4931 scope.go:117] "RemoveContainer" containerID="685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.294506 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.295120 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.316625 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.361346 4931 scope.go:117] "RemoveContainer" containerID="89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3" Nov 29 00:20:29 crc kubenswrapper[4931]: E1129 00:20:29.363933 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3\": container with ID starting with 89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3 not found: ID does not exist" containerID="89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.363996 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3"} err="failed to get container status \"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3\": rpc error: code = NotFound desc = could not find container \"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3\": container with ID starting with 89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3 not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.364047 4931 scope.go:117] "RemoveContainer" containerID="9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3" Nov 29 00:20:29 crc kubenswrapper[4931]: E1129 00:20:29.364358 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3\": container with ID starting with 9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3 not found: ID does not exist" containerID="9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.364420 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3"} err="failed to get container status \"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3\": rpc error: code = NotFound desc = could not find container \"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3\": container with ID starting with 9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3 not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.364457 4931 scope.go:117] "RemoveContainer" containerID="d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9" Nov 29 00:20:29 crc kubenswrapper[4931]: E1129 00:20:29.365081 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9\": container with ID starting with d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9 not found: ID does not exist" containerID="d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.365101 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9"} err="failed to get container status \"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9\": rpc error: code = NotFound desc = could not find container \"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9\": container with ID starting with d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9 not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.365114 4931 scope.go:117] "RemoveContainer" containerID="685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c" Nov 29 00:20:29 crc kubenswrapper[4931]: E1129 00:20:29.365280 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c\": container with ID starting with 685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c not found: ID does not exist" containerID="685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.365296 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c"} err="failed to get container status \"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c\": rpc error: code = NotFound desc = could not find container \"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c\": container with ID starting with 685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.365307 4931 scope.go:117] "RemoveContainer" containerID="89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.366365 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3"} err="failed to get container status \"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3\": rpc error: code = NotFound desc = could not find container \"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3\": container with ID starting with 89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3 not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.366392 4931 scope.go:117] "RemoveContainer" containerID="9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.367511 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3"} err="failed to get container status \"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3\": rpc error: code = NotFound desc = could not find container \"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3\": container with ID starting with 9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3 not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.367559 4931 scope.go:117] "RemoveContainer" containerID="d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.367942 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9"} err="failed to get container status \"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9\": rpc error: code = NotFound desc = could not find container \"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9\": container with ID starting with d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9 not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.368055 4931 scope.go:117] "RemoveContainer" containerID="685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.368747 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c"} err="failed to get container status \"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c\": rpc error: code = NotFound desc = could not find container \"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c\": container with ID starting with 685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.368769 4931 scope.go:117] "RemoveContainer" containerID="89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.369235 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3"} err="failed to get container status \"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3\": rpc error: code = NotFound desc = could not find container \"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3\": container with ID starting with 89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3 not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.369250 4931 scope.go:117] "RemoveContainer" containerID="9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.369284 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-config-data\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.369336 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-run-httpd\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.369358 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-scripts\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.369418 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.369474 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4p7p\" (UniqueName: \"kubernetes.io/projected/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-kube-api-access-z4p7p\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.369525 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-log-httpd\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.369612 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.370007 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3"} err="failed to get container status \"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3\": rpc error: code = NotFound desc = could not find container \"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3\": container with ID starting with 9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3 not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.370034 4931 scope.go:117] "RemoveContainer" containerID="d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.370756 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9"} err="failed to get container status \"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9\": rpc error: code = NotFound desc = could not find container \"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9\": container with ID starting with d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9 not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.370793 4931 scope.go:117] "RemoveContainer" containerID="685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.371128 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c"} err="failed to get container status \"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c\": rpc error: code = NotFound desc = could not find container \"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c\": container with ID starting with 685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.371159 4931 scope.go:117] "RemoveContainer" containerID="89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.371412 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3"} err="failed to get container status \"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3\": rpc error: code = NotFound desc = could not find container \"89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3\": container with ID starting with 89450bed44001301b88f967f906228194cc1a8b4985b3cdfc61d167c596a27c3 not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.371480 4931 scope.go:117] "RemoveContainer" containerID="9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.372387 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3"} err="failed to get container status \"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3\": rpc error: code = NotFound desc = could not find container \"9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3\": container with ID starting with 9808220d5c2f0f0250fa087a2610414e6e08516fcd5ed6d60675551b752dd1a3 not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.372416 4931 scope.go:117] "RemoveContainer" containerID="d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.372954 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9"} err="failed to get container status \"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9\": rpc error: code = NotFound desc = could not find container \"d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9\": container with ID starting with d78ee5057a4419cc0ded7c859386167d409c3b10bfe6938da97b4bfac24a99d9 not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.372995 4931 scope.go:117] "RemoveContainer" containerID="685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.373213 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c"} err="failed to get container status \"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c\": rpc error: code = NotFound desc = could not find container \"685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c\": container with ID starting with 685a4ac13b600cb59eaa533d461cecb67c6752e9a0e3a7db2ea63abed2a1289c not found: ID does not exist" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.471489 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4p7p\" (UniqueName: \"kubernetes.io/projected/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-kube-api-access-z4p7p\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.471568 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-log-httpd\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.471697 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.471763 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-config-data\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.471803 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-run-httpd\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.471884 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-scripts\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.471945 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.472128 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-log-httpd\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.473015 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-run-httpd\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.476828 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.477086 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-scripts\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.482477 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-config-data\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.484660 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.492104 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4p7p\" (UniqueName: \"kubernetes.io/projected/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-kube-api-access-z4p7p\") pod \"ceilometer-0\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: W1129 00:20:29.496353 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda52f3c1d_5aa8_4c5c_a51b_71b29ea5e0f5.slice/crio-cd73822c697c0cb18d896dcdbaa817d215e836dc0e293b82fe77a0e3f380aa0e WatchSource:0}: Error finding container cd73822c697c0cb18d896dcdbaa817d215e836dc0e293b82fe77a0e3f380aa0e: Status 404 returned error can't find the container with id cd73822c697c0cb18d896dcdbaa817d215e836dc0e293b82fe77a0e3f380aa0e Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.501265 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dw4x9"] Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.598508 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-68b1-account-create-update-4flsl"] Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.608154 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3603-account-create-update-8rgkv"] Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.612036 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:20:29 crc kubenswrapper[4931]: W1129 00:20:29.682068 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46687ab2_074a_4007_b89a_a962c9b1d14d.slice/crio-966d26054f046d34fe79fdc5fa3f9b9f5ace18e7cc8764fb30380ef39b2876a3 WatchSource:0}: Error finding container 966d26054f046d34fe79fdc5fa3f9b9f5ace18e7cc8764fb30380ef39b2876a3: Status 404 returned error can't find the container with id 966d26054f046d34fe79fdc5fa3f9b9f5ace18e7cc8764fb30380ef39b2876a3 Nov 29 00:20:29 crc kubenswrapper[4931]: W1129 00:20:29.684182 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda57e64d1_340e_4e80_8112_16625e6d7d6e.slice/crio-8ffe52acf4e6ea3fd90a85beb9a5d15d7fea84f4a8d8a18b3d8c46c33d54f282 WatchSource:0}: Error finding container 8ffe52acf4e6ea3fd90a85beb9a5d15d7fea84f4a8d8a18b3d8c46c33d54f282: Status 404 returned error can't find the container with id 8ffe52acf4e6ea3fd90a85beb9a5d15d7fea84f4a8d8a18b3d8c46c33d54f282 Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.727836 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4b32-account-create-update-nhrbj"] Nov 29 00:20:29 crc kubenswrapper[4931]: W1129 00:20:29.761994 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f82d911_133a_4c7a_87a4_9684bb066cc0.slice/crio-5984dfb3dc0ec21b97c4de2cae1c11457dd0e839e545f97f21afd31c8baf3914 WatchSource:0}: Error finding container 5984dfb3dc0ec21b97c4de2cae1c11457dd0e839e545f97f21afd31c8baf3914: Status 404 returned error can't find the container with id 5984dfb3dc0ec21b97c4de2cae1c11457dd0e839e545f97f21afd31c8baf3914 Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.863338 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-68b1-account-create-update-4flsl" event={"ID":"46687ab2-074a-4007-b89a-a962c9b1d14d","Type":"ContainerStarted","Data":"966d26054f046d34fe79fdc5fa3f9b9f5ace18e7cc8764fb30380ef39b2876a3"} Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.872529 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3603-account-create-update-8rgkv" event={"ID":"a57e64d1-340e-4e80-8112-16625e6d7d6e","Type":"ContainerStarted","Data":"8ffe52acf4e6ea3fd90a85beb9a5d15d7fea84f4a8d8a18b3d8c46c33d54f282"} Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.879438 4931 generic.go:334] "Generic (PLEG): container finished" podID="2f6ca58a-4bc4-4277-891d-db214595cb56" containerID="a61489bacc4a02782d504b27decc932ac85e50c8676418a8d3d2613218a52f0d" exitCode=0 Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.879502 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-r7n7s" event={"ID":"2f6ca58a-4bc4-4277-891d-db214595cb56","Type":"ContainerDied","Data":"a61489bacc4a02782d504b27decc932ac85e50c8676418a8d3d2613218a52f0d"} Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.879521 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-r7n7s" event={"ID":"2f6ca58a-4bc4-4277-891d-db214595cb56","Type":"ContainerStarted","Data":"aa4c34849eee464084f3295a4d855424ccb7c217f6ff62bb5799afc93e3dc159"} Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.881848 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dw4x9" event={"ID":"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5","Type":"ContainerStarted","Data":"69bf7ea057acdfc9d180e8d17d962cc2b048ebda75fe1998c51db15f04dce044"} Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.881866 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dw4x9" event={"ID":"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5","Type":"ContainerStarted","Data":"cd73822c697c0cb18d896dcdbaa817d215e836dc0e293b82fe77a0e3f380aa0e"} Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.883791 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" event={"ID":"5f82d911-133a-4c7a-87a4-9684bb066cc0","Type":"ContainerStarted","Data":"5984dfb3dc0ec21b97c4de2cae1c11457dd0e839e545f97f21afd31c8baf3914"} Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.886391 4931 generic.go:334] "Generic (PLEG): container finished" podID="6284f70c-aaa2-4182-98fd-82fb59295e1f" containerID="d50ca037edd2752ec5e0926af44bb5d8cead4abc1e1ad1a96b51f1b117b215d1" exitCode=0 Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.886417 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n899b" event={"ID":"6284f70c-aaa2-4182-98fd-82fb59295e1f","Type":"ContainerDied","Data":"d50ca037edd2752ec5e0926af44bb5d8cead4abc1e1ad1a96b51f1b117b215d1"} Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.886532 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n899b" event={"ID":"6284f70c-aaa2-4182-98fd-82fb59295e1f","Type":"ContainerStarted","Data":"32338bb42aefbb48634e1a7cfcfd848b335e7c1f8b87666f6a84ffcc4fc05706"} Nov 29 00:20:29 crc kubenswrapper[4931]: I1129 00:20:29.915250 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-dw4x9" podStartSLOduration=1.915232505 podStartE2EDuration="1.915232505s" podCreationTimestamp="2025-11-29 00:20:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:20:29.908706586 +0000 UTC m=+1243.070599818" watchObservedRunningTime="2025-11-29 00:20:29.915232505 +0000 UTC m=+1243.077125737" Nov 29 00:20:30 crc kubenswrapper[4931]: I1129 00:20:30.083308 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:20:30 crc kubenswrapper[4931]: W1129 00:20:30.100322 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode52b8c44_4758_44ec_ac4d_2f0e5738fd2f.slice/crio-a7358132857055de9332bc8e8d72bb285454ca4ef11c7b705006b4816ed01200 WatchSource:0}: Error finding container a7358132857055de9332bc8e8d72bb285454ca4ef11c7b705006b4816ed01200: Status 404 returned error can't find the container with id a7358132857055de9332bc8e8d72bb285454ca4ef11c7b705006b4816ed01200 Nov 29 00:20:30 crc kubenswrapper[4931]: I1129 00:20:30.898463 4931 generic.go:334] "Generic (PLEG): container finished" podID="5f82d911-133a-4c7a-87a4-9684bb066cc0" containerID="d89c36e7e0927e0ca927465aaa61f9848057756bee2cd62e2c72bc146aa7e137" exitCode=0 Nov 29 00:20:30 crc kubenswrapper[4931]: I1129 00:20:30.898978 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" event={"ID":"5f82d911-133a-4c7a-87a4-9684bb066cc0","Type":"ContainerDied","Data":"d89c36e7e0927e0ca927465aaa61f9848057756bee2cd62e2c72bc146aa7e137"} Nov 29 00:20:30 crc kubenswrapper[4931]: I1129 00:20:30.902870 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f","Type":"ContainerStarted","Data":"a7358132857055de9332bc8e8d72bb285454ca4ef11c7b705006b4816ed01200"} Nov 29 00:20:30 crc kubenswrapper[4931]: I1129 00:20:30.906754 4931 generic.go:334] "Generic (PLEG): container finished" podID="46687ab2-074a-4007-b89a-a962c9b1d14d" containerID="f8eda5ccdc2aede73e4bc9eaab88cdd9f2ca4a37c25f35e76630e432175ea07b" exitCode=0 Nov 29 00:20:30 crc kubenswrapper[4931]: I1129 00:20:30.906851 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-68b1-account-create-update-4flsl" event={"ID":"46687ab2-074a-4007-b89a-a962c9b1d14d","Type":"ContainerDied","Data":"f8eda5ccdc2aede73e4bc9eaab88cdd9f2ca4a37c25f35e76630e432175ea07b"} Nov 29 00:20:30 crc kubenswrapper[4931]: I1129 00:20:30.908044 4931 generic.go:334] "Generic (PLEG): container finished" podID="a57e64d1-340e-4e80-8112-16625e6d7d6e" containerID="0489f9f59e6d194d1f2a44f58a42656cb27237e5e71d01d367eba8c79ebc2c70" exitCode=0 Nov 29 00:20:30 crc kubenswrapper[4931]: I1129 00:20:30.908094 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3603-account-create-update-8rgkv" event={"ID":"a57e64d1-340e-4e80-8112-16625e6d7d6e","Type":"ContainerDied","Data":"0489f9f59e6d194d1f2a44f58a42656cb27237e5e71d01d367eba8c79ebc2c70"} Nov 29 00:20:30 crc kubenswrapper[4931]: I1129 00:20:30.910687 4931 generic.go:334] "Generic (PLEG): container finished" podID="a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5" containerID="69bf7ea057acdfc9d180e8d17d962cc2b048ebda75fe1998c51db15f04dce044" exitCode=0 Nov 29 00:20:30 crc kubenswrapper[4931]: I1129 00:20:30.910783 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dw4x9" event={"ID":"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5","Type":"ContainerDied","Data":"69bf7ea057acdfc9d180e8d17d962cc2b048ebda75fe1998c51db15f04dce044"} Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.225920 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79df5ada-b426-413e-a995-287294286bed" path="/var/lib/kubelet/pods/79df5ada-b426-413e-a995-287294286bed/volumes" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.361929 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n899b" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.369240 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-r7n7s" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.514193 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6284f70c-aaa2-4182-98fd-82fb59295e1f-operator-scripts\") pod \"6284f70c-aaa2-4182-98fd-82fb59295e1f\" (UID: \"6284f70c-aaa2-4182-98fd-82fb59295e1f\") " Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.514274 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc9l9\" (UniqueName: \"kubernetes.io/projected/6284f70c-aaa2-4182-98fd-82fb59295e1f-kube-api-access-mc9l9\") pod \"6284f70c-aaa2-4182-98fd-82fb59295e1f\" (UID: \"6284f70c-aaa2-4182-98fd-82fb59295e1f\") " Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.514366 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f6ca58a-4bc4-4277-891d-db214595cb56-operator-scripts\") pod \"2f6ca58a-4bc4-4277-891d-db214595cb56\" (UID: \"2f6ca58a-4bc4-4277-891d-db214595cb56\") " Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.514407 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t5hv\" (UniqueName: \"kubernetes.io/projected/2f6ca58a-4bc4-4277-891d-db214595cb56-kube-api-access-7t5hv\") pod \"2f6ca58a-4bc4-4277-891d-db214595cb56\" (UID: \"2f6ca58a-4bc4-4277-891d-db214595cb56\") " Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.517555 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6284f70c-aaa2-4182-98fd-82fb59295e1f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6284f70c-aaa2-4182-98fd-82fb59295e1f" (UID: "6284f70c-aaa2-4182-98fd-82fb59295e1f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.518090 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f6ca58a-4bc4-4277-891d-db214595cb56-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2f6ca58a-4bc4-4277-891d-db214595cb56" (UID: "2f6ca58a-4bc4-4277-891d-db214595cb56"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.520544 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f6ca58a-4bc4-4277-891d-db214595cb56-kube-api-access-7t5hv" (OuterVolumeSpecName: "kube-api-access-7t5hv") pod "2f6ca58a-4bc4-4277-891d-db214595cb56" (UID: "2f6ca58a-4bc4-4277-891d-db214595cb56"). InnerVolumeSpecName "kube-api-access-7t5hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.532352 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6284f70c-aaa2-4182-98fd-82fb59295e1f-kube-api-access-mc9l9" (OuterVolumeSpecName: "kube-api-access-mc9l9") pod "6284f70c-aaa2-4182-98fd-82fb59295e1f" (UID: "6284f70c-aaa2-4182-98fd-82fb59295e1f"). InnerVolumeSpecName "kube-api-access-mc9l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.617141 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6284f70c-aaa2-4182-98fd-82fb59295e1f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.617185 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc9l9\" (UniqueName: \"kubernetes.io/projected/6284f70c-aaa2-4182-98fd-82fb59295e1f-kube-api-access-mc9l9\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.617202 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f6ca58a-4bc4-4277-891d-db214595cb56-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.617215 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t5hv\" (UniqueName: \"kubernetes.io/projected/2f6ca58a-4bc4-4277-891d-db214595cb56-kube-api-access-7t5hv\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.920120 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f","Type":"ContainerStarted","Data":"02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d"} Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.922726 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-r7n7s" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.922838 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-r7n7s" event={"ID":"2f6ca58a-4bc4-4277-891d-db214595cb56","Type":"ContainerDied","Data":"aa4c34849eee464084f3295a4d855424ccb7c217f6ff62bb5799afc93e3dc159"} Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.922902 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa4c34849eee464084f3295a4d855424ccb7c217f6ff62bb5799afc93e3dc159" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.925649 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n899b" Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.929658 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n899b" event={"ID":"6284f70c-aaa2-4182-98fd-82fb59295e1f","Type":"ContainerDied","Data":"32338bb42aefbb48634e1a7cfcfd848b335e7c1f8b87666f6a84ffcc4fc05706"} Nov 29 00:20:31 crc kubenswrapper[4931]: I1129 00:20:31.929703 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32338bb42aefbb48634e1a7cfcfd848b335e7c1f8b87666f6a84ffcc4fc05706" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.150042 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.150125 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.179848 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.220417 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.374283 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dw4x9" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.451043 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5-operator-scripts\") pod \"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5\" (UID: \"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5\") " Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.451109 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5chdf\" (UniqueName: \"kubernetes.io/projected/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5-kube-api-access-5chdf\") pod \"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5\" (UID: \"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5\") " Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.451522 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5" (UID: "a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.457353 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5-kube-api-access-5chdf" (OuterVolumeSpecName: "kube-api-access-5chdf") pod "a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5" (UID: "a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5"). InnerVolumeSpecName "kube-api-access-5chdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.502415 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3603-account-create-update-8rgkv" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.510211 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-68b1-account-create-update-4flsl" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.517435 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.553609 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.553639 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5chdf\" (UniqueName: \"kubernetes.io/projected/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5-kube-api-access-5chdf\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.654245 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a57e64d1-340e-4e80-8112-16625e6d7d6e-operator-scripts\") pod \"a57e64d1-340e-4e80-8112-16625e6d7d6e\" (UID: \"a57e64d1-340e-4e80-8112-16625e6d7d6e\") " Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.654333 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6zdh\" (UniqueName: \"kubernetes.io/projected/a57e64d1-340e-4e80-8112-16625e6d7d6e-kube-api-access-x6zdh\") pod \"a57e64d1-340e-4e80-8112-16625e6d7d6e\" (UID: \"a57e64d1-340e-4e80-8112-16625e6d7d6e\") " Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.654358 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46687ab2-074a-4007-b89a-a962c9b1d14d-operator-scripts\") pod \"46687ab2-074a-4007-b89a-a962c9b1d14d\" (UID: \"46687ab2-074a-4007-b89a-a962c9b1d14d\") " Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.654401 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f82d911-133a-4c7a-87a4-9684bb066cc0-operator-scripts\") pod \"5f82d911-133a-4c7a-87a4-9684bb066cc0\" (UID: \"5f82d911-133a-4c7a-87a4-9684bb066cc0\") " Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.654425 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g4kg\" (UniqueName: \"kubernetes.io/projected/5f82d911-133a-4c7a-87a4-9684bb066cc0-kube-api-access-5g4kg\") pod \"5f82d911-133a-4c7a-87a4-9684bb066cc0\" (UID: \"5f82d911-133a-4c7a-87a4-9684bb066cc0\") " Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.654513 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7tqx\" (UniqueName: \"kubernetes.io/projected/46687ab2-074a-4007-b89a-a962c9b1d14d-kube-api-access-h7tqx\") pod \"46687ab2-074a-4007-b89a-a962c9b1d14d\" (UID: \"46687ab2-074a-4007-b89a-a962c9b1d14d\") " Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.658838 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a57e64d1-340e-4e80-8112-16625e6d7d6e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a57e64d1-340e-4e80-8112-16625e6d7d6e" (UID: "a57e64d1-340e-4e80-8112-16625e6d7d6e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.658908 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46687ab2-074a-4007-b89a-a962c9b1d14d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "46687ab2-074a-4007-b89a-a962c9b1d14d" (UID: "46687ab2-074a-4007-b89a-a962c9b1d14d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.659179 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f82d911-133a-4c7a-87a4-9684bb066cc0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5f82d911-133a-4c7a-87a4-9684bb066cc0" (UID: "5f82d911-133a-4c7a-87a4-9684bb066cc0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.659281 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46687ab2-074a-4007-b89a-a962c9b1d14d-kube-api-access-h7tqx" (OuterVolumeSpecName: "kube-api-access-h7tqx") pod "46687ab2-074a-4007-b89a-a962c9b1d14d" (UID: "46687ab2-074a-4007-b89a-a962c9b1d14d"). InnerVolumeSpecName "kube-api-access-h7tqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.663263 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f82d911-133a-4c7a-87a4-9684bb066cc0-kube-api-access-5g4kg" (OuterVolumeSpecName: "kube-api-access-5g4kg") pod "5f82d911-133a-4c7a-87a4-9684bb066cc0" (UID: "5f82d911-133a-4c7a-87a4-9684bb066cc0"). InnerVolumeSpecName "kube-api-access-5g4kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.663326 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a57e64d1-340e-4e80-8112-16625e6d7d6e-kube-api-access-x6zdh" (OuterVolumeSpecName: "kube-api-access-x6zdh") pod "a57e64d1-340e-4e80-8112-16625e6d7d6e" (UID: "a57e64d1-340e-4e80-8112-16625e6d7d6e"). InnerVolumeSpecName "kube-api-access-x6zdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.756885 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a57e64d1-340e-4e80-8112-16625e6d7d6e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.756933 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6zdh\" (UniqueName: \"kubernetes.io/projected/a57e64d1-340e-4e80-8112-16625e6d7d6e-kube-api-access-x6zdh\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.756951 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46687ab2-074a-4007-b89a-a962c9b1d14d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.756960 4931 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f82d911-133a-4c7a-87a4-9684bb066cc0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.756968 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g4kg\" (UniqueName: \"kubernetes.io/projected/5f82d911-133a-4c7a-87a4-9684bb066cc0-kube-api-access-5g4kg\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.756976 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7tqx\" (UniqueName: \"kubernetes.io/projected/46687ab2-074a-4007-b89a-a962c9b1d14d-kube-api-access-h7tqx\") on node \"crc\" DevicePath \"\"" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.936903 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-68b1-account-create-update-4flsl" event={"ID":"46687ab2-074a-4007-b89a-a962c9b1d14d","Type":"ContainerDied","Data":"966d26054f046d34fe79fdc5fa3f9b9f5ace18e7cc8764fb30380ef39b2876a3"} Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.937303 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="966d26054f046d34fe79fdc5fa3f9b9f5ace18e7cc8764fb30380ef39b2876a3" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.936924 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-68b1-account-create-update-4flsl" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.938397 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3603-account-create-update-8rgkv" event={"ID":"a57e64d1-340e-4e80-8112-16625e6d7d6e","Type":"ContainerDied","Data":"8ffe52acf4e6ea3fd90a85beb9a5d15d7fea84f4a8d8a18b3d8c46c33d54f282"} Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.938441 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ffe52acf4e6ea3fd90a85beb9a5d15d7fea84f4a8d8a18b3d8c46c33d54f282" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.938503 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3603-account-create-update-8rgkv" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.944938 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dw4x9" event={"ID":"a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5","Type":"ContainerDied","Data":"cd73822c697c0cb18d896dcdbaa817d215e836dc0e293b82fe77a0e3f380aa0e"} Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.944971 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd73822c697c0cb18d896dcdbaa817d215e836dc0e293b82fe77a0e3f380aa0e" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.945036 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dw4x9" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.947744 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.947739 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4b32-account-create-update-nhrbj" event={"ID":"5f82d911-133a-4c7a-87a4-9684bb066cc0","Type":"ContainerDied","Data":"5984dfb3dc0ec21b97c4de2cae1c11457dd0e839e545f97f21afd31c8baf3914"} Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.947900 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5984dfb3dc0ec21b97c4de2cae1c11457dd0e839e545f97f21afd31c8baf3914" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.949314 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f","Type":"ContainerStarted","Data":"bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66"} Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.949860 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 29 00:20:32 crc kubenswrapper[4931]: I1129 00:20:32.949903 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 29 00:20:33 crc kubenswrapper[4931]: I1129 00:20:33.254041 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:33 crc kubenswrapper[4931]: I1129 00:20:33.256152 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:33 crc kubenswrapper[4931]: I1129 00:20:33.298108 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:33 crc kubenswrapper[4931]: I1129 00:20:33.313081 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:33 crc kubenswrapper[4931]: I1129 00:20:33.959742 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f","Type":"ContainerStarted","Data":"fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50"} Nov 29 00:20:33 crc kubenswrapper[4931]: I1129 00:20:33.960360 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:33 crc kubenswrapper[4931]: I1129 00:20:33.960392 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:34 crc kubenswrapper[4931]: I1129 00:20:34.947504 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 29 00:20:34 crc kubenswrapper[4931]: I1129 00:20:34.947948 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 29 00:20:35 crc kubenswrapper[4931]: I1129 00:20:35.900758 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:35 crc kubenswrapper[4931]: I1129 00:20:35.904051 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 29 00:20:35 crc kubenswrapper[4931]: I1129 00:20:35.979407 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f","Type":"ContainerStarted","Data":"d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb"} Nov 29 00:20:35 crc kubenswrapper[4931]: I1129 00:20:35.979730 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 00:20:36 crc kubenswrapper[4931]: I1129 00:20:36.013966 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.081654397 podStartE2EDuration="7.013940709s" podCreationTimestamp="2025-11-29 00:20:29 +0000 UTC" firstStartedPulling="2025-11-29 00:20:30.102720315 +0000 UTC m=+1243.264613547" lastFinishedPulling="2025-11-29 00:20:35.035006627 +0000 UTC m=+1248.196899859" observedRunningTime="2025-11-29 00:20:36.009030586 +0000 UTC m=+1249.170923818" watchObservedRunningTime="2025-11-29 00:20:36.013940709 +0000 UTC m=+1249.175833941" Nov 29 00:20:36 crc kubenswrapper[4931]: I1129 00:20:36.049668 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:20:36 crc kubenswrapper[4931]: E1129 00:20:36.051628 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:20:36 crc kubenswrapper[4931]: E1129 00:20:36.051681 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-764cb44577-7q67f: configmap "swift-ring-files" not found Nov 29 00:20:36 crc kubenswrapper[4931]: E1129 00:20:36.052154 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift podName:550e16f7-f101-4a91-8e2f-45af3430d6b9 nodeName:}" failed. No retries permitted until 2025-11-29 00:21:08.052132915 +0000 UTC m=+1281.214026147 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift") pod "swift-proxy-764cb44577-7q67f" (UID: "550e16f7-f101-4a91-8e2f-45af3430d6b9") : configmap "swift-ring-files" not found Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.651673 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b4dlr"] Nov 29 00:20:38 crc kubenswrapper[4931]: E1129 00:20:38.652413 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46687ab2-074a-4007-b89a-a962c9b1d14d" containerName="mariadb-account-create-update" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.652426 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="46687ab2-074a-4007-b89a-a962c9b1d14d" containerName="mariadb-account-create-update" Nov 29 00:20:38 crc kubenswrapper[4931]: E1129 00:20:38.652447 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f82d911-133a-4c7a-87a4-9684bb066cc0" containerName="mariadb-account-create-update" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.652453 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f82d911-133a-4c7a-87a4-9684bb066cc0" containerName="mariadb-account-create-update" Nov 29 00:20:38 crc kubenswrapper[4931]: E1129 00:20:38.652464 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57e64d1-340e-4e80-8112-16625e6d7d6e" containerName="mariadb-account-create-update" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.652470 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57e64d1-340e-4e80-8112-16625e6d7d6e" containerName="mariadb-account-create-update" Nov 29 00:20:38 crc kubenswrapper[4931]: E1129 00:20:38.652485 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f6ca58a-4bc4-4277-891d-db214595cb56" containerName="mariadb-database-create" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.652490 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f6ca58a-4bc4-4277-891d-db214595cb56" containerName="mariadb-database-create" Nov 29 00:20:38 crc kubenswrapper[4931]: E1129 00:20:38.652501 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5" containerName="mariadb-database-create" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.652507 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5" containerName="mariadb-database-create" Nov 29 00:20:38 crc kubenswrapper[4931]: E1129 00:20:38.652524 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6284f70c-aaa2-4182-98fd-82fb59295e1f" containerName="mariadb-database-create" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.652530 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="6284f70c-aaa2-4182-98fd-82fb59295e1f" containerName="mariadb-database-create" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.652684 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="6284f70c-aaa2-4182-98fd-82fb59295e1f" containerName="mariadb-database-create" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.652695 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f6ca58a-4bc4-4277-891d-db214595cb56" containerName="mariadb-database-create" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.652708 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="46687ab2-074a-4007-b89a-a962c9b1d14d" containerName="mariadb-account-create-update" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.652719 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f82d911-133a-4c7a-87a4-9684bb066cc0" containerName="mariadb-account-create-update" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.652728 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5" containerName="mariadb-database-create" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.652737 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a57e64d1-340e-4e80-8112-16625e6d7d6e" containerName="mariadb-account-create-update" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.653309 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.659278 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.659870 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bjxfs" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.660910 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.669544 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b4dlr"] Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.704290 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-config-data\") pod \"nova-cell0-conductor-db-sync-b4dlr\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.704412 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-b4dlr\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.704476 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-scripts\") pod \"nova-cell0-conductor-db-sync-b4dlr\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.704608 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m2jn\" (UniqueName: \"kubernetes.io/projected/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-kube-api-access-6m2jn\") pod \"nova-cell0-conductor-db-sync-b4dlr\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.806188 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-config-data\") pod \"nova-cell0-conductor-db-sync-b4dlr\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.806297 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-b4dlr\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.806353 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-scripts\") pod \"nova-cell0-conductor-db-sync-b4dlr\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.806428 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m2jn\" (UniqueName: \"kubernetes.io/projected/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-kube-api-access-6m2jn\") pod \"nova-cell0-conductor-db-sync-b4dlr\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.812062 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-scripts\") pod \"nova-cell0-conductor-db-sync-b4dlr\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.819488 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-b4dlr\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.822580 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-config-data\") pod \"nova-cell0-conductor-db-sync-b4dlr\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.823563 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m2jn\" (UniqueName: \"kubernetes.io/projected/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-kube-api-access-6m2jn\") pod \"nova-cell0-conductor-db-sync-b4dlr\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:38 crc kubenswrapper[4931]: I1129 00:20:38.973084 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:20:39 crc kubenswrapper[4931]: I1129 00:20:39.418109 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b4dlr"] Nov 29 00:20:40 crc kubenswrapper[4931]: I1129 00:20:40.012344 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b4dlr" event={"ID":"f40ca0f5-3fa9-4003-87cb-aed2feeaf545","Type":"ContainerStarted","Data":"60056f1c906738cb210646089d0e26e137734e7ffe5eff4c98f9d54e9e78031a"} Nov 29 00:20:48 crc kubenswrapper[4931]: I1129 00:20:48.092734 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b4dlr" event={"ID":"f40ca0f5-3fa9-4003-87cb-aed2feeaf545","Type":"ContainerStarted","Data":"69a758d0710ba7c61b3fa8ff23863c7140f3896b1f2832cdf1af16162b8bc156"} Nov 29 00:20:48 crc kubenswrapper[4931]: I1129 00:20:48.115452 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-b4dlr" podStartSLOduration=2.246169005 podStartE2EDuration="10.113357689s" podCreationTimestamp="2025-11-29 00:20:38 +0000 UTC" firstStartedPulling="2025-11-29 00:20:39.423301212 +0000 UTC m=+1252.585194444" lastFinishedPulling="2025-11-29 00:20:47.290489896 +0000 UTC m=+1260.452383128" observedRunningTime="2025-11-29 00:20:48.109161097 +0000 UTC m=+1261.271054339" watchObservedRunningTime="2025-11-29 00:20:48.113357689 +0000 UTC m=+1261.275250961" Nov 29 00:20:53 crc kubenswrapper[4931]: I1129 00:20:53.734668 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:20:53 crc kubenswrapper[4931]: I1129 00:20:53.736424 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:20:53 crc kubenswrapper[4931]: I1129 00:20:53.736585 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:20:53 crc kubenswrapper[4931]: I1129 00:20:53.737447 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7ca6926ea091667ee0b6b53772fafdc3830c2a5a094ad559af9957d016e82d86"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 00:20:53 crc kubenswrapper[4931]: I1129 00:20:53.737617 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://7ca6926ea091667ee0b6b53772fafdc3830c2a5a094ad559af9957d016e82d86" gracePeriod=600 Nov 29 00:20:54 crc kubenswrapper[4931]: I1129 00:20:54.156706 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="7ca6926ea091667ee0b6b53772fafdc3830c2a5a094ad559af9957d016e82d86" exitCode=0 Nov 29 00:20:54 crc kubenswrapper[4931]: I1129 00:20:54.156779 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"7ca6926ea091667ee0b6b53772fafdc3830c2a5a094ad559af9957d016e82d86"} Nov 29 00:20:54 crc kubenswrapper[4931]: I1129 00:20:54.157067 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"e130783d6b1a728f7dced4948fd5956d2a8b18dc3e74e1cf5f76c0244465eeba"} Nov 29 00:20:54 crc kubenswrapper[4931]: I1129 00:20:54.157090 4931 scope.go:117] "RemoveContainer" containerID="63e24d3607eb4b6f4826de9209c8d28131ebf746bafba6dd38a837ce70168944" Nov 29 00:20:59 crc kubenswrapper[4931]: I1129 00:20:59.619207 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 29 00:21:01 crc kubenswrapper[4931]: I1129 00:21:01.230295 4931 generic.go:334] "Generic (PLEG): container finished" podID="f40ca0f5-3fa9-4003-87cb-aed2feeaf545" containerID="69a758d0710ba7c61b3fa8ff23863c7140f3896b1f2832cdf1af16162b8bc156" exitCode=0 Nov 29 00:21:01 crc kubenswrapper[4931]: I1129 00:21:01.230350 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b4dlr" event={"ID":"f40ca0f5-3fa9-4003-87cb-aed2feeaf545","Type":"ContainerDied","Data":"69a758d0710ba7c61b3fa8ff23863c7140f3896b1f2832cdf1af16162b8bc156"} Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.588081 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.688990 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m2jn\" (UniqueName: \"kubernetes.io/projected/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-kube-api-access-6m2jn\") pod \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.689035 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-config-data\") pod \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.689078 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-combined-ca-bundle\") pod \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.689826 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-scripts\") pod \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\" (UID: \"f40ca0f5-3fa9-4003-87cb-aed2feeaf545\") " Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.694590 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-kube-api-access-6m2jn" (OuterVolumeSpecName: "kube-api-access-6m2jn") pod "f40ca0f5-3fa9-4003-87cb-aed2feeaf545" (UID: "f40ca0f5-3fa9-4003-87cb-aed2feeaf545"). InnerVolumeSpecName "kube-api-access-6m2jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.694966 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-scripts" (OuterVolumeSpecName: "scripts") pod "f40ca0f5-3fa9-4003-87cb-aed2feeaf545" (UID: "f40ca0f5-3fa9-4003-87cb-aed2feeaf545"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.714161 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-config-data" (OuterVolumeSpecName: "config-data") pod "f40ca0f5-3fa9-4003-87cb-aed2feeaf545" (UID: "f40ca0f5-3fa9-4003-87cb-aed2feeaf545"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.715598 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f40ca0f5-3fa9-4003-87cb-aed2feeaf545" (UID: "f40ca0f5-3fa9-4003-87cb-aed2feeaf545"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.792299 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m2jn\" (UniqueName: \"kubernetes.io/projected/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-kube-api-access-6m2jn\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.792730 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.792791 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:02 crc kubenswrapper[4931]: I1129 00:21:02.792844 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40ca0f5-3fa9-4003-87cb-aed2feeaf545-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.059779 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.060117 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="be472357-f01e-4e83-8903-49c0e2f60c97" containerName="kube-state-metrics" containerID="cri-o://456cc2174afa59dc77de92827cdb152c0ca1ed767f352bf123447d6dd0c91c71" gracePeriod=30 Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.253494 4931 generic.go:334] "Generic (PLEG): container finished" podID="be472357-f01e-4e83-8903-49c0e2f60c97" containerID="456cc2174afa59dc77de92827cdb152c0ca1ed767f352bf123447d6dd0c91c71" exitCode=2 Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.253781 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"be472357-f01e-4e83-8903-49c0e2f60c97","Type":"ContainerDied","Data":"456cc2174afa59dc77de92827cdb152c0ca1ed767f352bf123447d6dd0c91c71"} Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.255674 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b4dlr" event={"ID":"f40ca0f5-3fa9-4003-87cb-aed2feeaf545","Type":"ContainerDied","Data":"60056f1c906738cb210646089d0e26e137734e7ffe5eff4c98f9d54e9e78031a"} Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.255707 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60056f1c906738cb210646089d0e26e137734e7ffe5eff4c98f9d54e9e78031a" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.255748 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b4dlr" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.347827 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 29 00:21:03 crc kubenswrapper[4931]: E1129 00:21:03.350304 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40ca0f5-3fa9-4003-87cb-aed2feeaf545" containerName="nova-cell0-conductor-db-sync" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.350323 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40ca0f5-3fa9-4003-87cb-aed2feeaf545" containerName="nova-cell0-conductor-db-sync" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.350575 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40ca0f5-3fa9-4003-87cb-aed2feeaf545" containerName="nova-cell0-conductor-db-sync" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.351322 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.356433 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.356788 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bjxfs" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.361384 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.408114 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dcc4f63-f88c-436a-96a6-cae43c016e0c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2dcc4f63-f88c-436a-96a6-cae43c016e0c\") " pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.408255 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dcc4f63-f88c-436a-96a6-cae43c016e0c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2dcc4f63-f88c-436a-96a6-cae43c016e0c\") " pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.408328 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxml4\" (UniqueName: \"kubernetes.io/projected/2dcc4f63-f88c-436a-96a6-cae43c016e0c-kube-api-access-xxml4\") pod \"nova-cell0-conductor-0\" (UID: \"2dcc4f63-f88c-436a-96a6-cae43c016e0c\") " pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.465939 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.509352 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsvwl\" (UniqueName: \"kubernetes.io/projected/be472357-f01e-4e83-8903-49c0e2f60c97-kube-api-access-qsvwl\") pod \"be472357-f01e-4e83-8903-49c0e2f60c97\" (UID: \"be472357-f01e-4e83-8903-49c0e2f60c97\") " Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.509694 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxml4\" (UniqueName: \"kubernetes.io/projected/2dcc4f63-f88c-436a-96a6-cae43c016e0c-kube-api-access-xxml4\") pod \"nova-cell0-conductor-0\" (UID: \"2dcc4f63-f88c-436a-96a6-cae43c016e0c\") " pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.509747 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dcc4f63-f88c-436a-96a6-cae43c016e0c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2dcc4f63-f88c-436a-96a6-cae43c016e0c\") " pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.509863 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dcc4f63-f88c-436a-96a6-cae43c016e0c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2dcc4f63-f88c-436a-96a6-cae43c016e0c\") " pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.514980 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dcc4f63-f88c-436a-96a6-cae43c016e0c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2dcc4f63-f88c-436a-96a6-cae43c016e0c\") " pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.515622 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dcc4f63-f88c-436a-96a6-cae43c016e0c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2dcc4f63-f88c-436a-96a6-cae43c016e0c\") " pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.515737 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be472357-f01e-4e83-8903-49c0e2f60c97-kube-api-access-qsvwl" (OuterVolumeSpecName: "kube-api-access-qsvwl") pod "be472357-f01e-4e83-8903-49c0e2f60c97" (UID: "be472357-f01e-4e83-8903-49c0e2f60c97"). InnerVolumeSpecName "kube-api-access-qsvwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.524506 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxml4\" (UniqueName: \"kubernetes.io/projected/2dcc4f63-f88c-436a-96a6-cae43c016e0c-kube-api-access-xxml4\") pod \"nova-cell0-conductor-0\" (UID: \"2dcc4f63-f88c-436a-96a6-cae43c016e0c\") " pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.610972 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsvwl\" (UniqueName: \"kubernetes.io/projected/be472357-f01e-4e83-8903-49c0e2f60c97-kube-api-access-qsvwl\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.697346 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:03 crc kubenswrapper[4931]: I1129 00:21:03.981480 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.266614 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2dcc4f63-f88c-436a-96a6-cae43c016e0c","Type":"ContainerStarted","Data":"9e2111d1adf4f2d8552a569f258ad4ecd122ec9ad2f86d98102454a76358c4a1"} Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.267050 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2dcc4f63-f88c-436a-96a6-cae43c016e0c","Type":"ContainerStarted","Data":"731a10e80e5f132818a9f7a458aa08c47dc0917547bedbea15a965f337d7fb86"} Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.268502 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.271838 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"be472357-f01e-4e83-8903-49c0e2f60c97","Type":"ContainerDied","Data":"1361fab9af24f7ae14afba21dcb17eadfe875b9d0ecaaefdc540e541f7e313ef"} Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.271877 4931 scope.go:117] "RemoveContainer" containerID="456cc2174afa59dc77de92827cdb152c0ca1ed767f352bf123447d6dd0c91c71" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.271984 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.289451 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.289432748 podStartE2EDuration="1.289432748s" podCreationTimestamp="2025-11-29 00:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:21:04.28640418 +0000 UTC m=+1277.448297432" watchObservedRunningTime="2025-11-29 00:21:04.289432748 +0000 UTC m=+1277.451325990" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.312432 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.322675 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.334245 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 00:21:04 crc kubenswrapper[4931]: E1129 00:21:04.334732 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be472357-f01e-4e83-8903-49c0e2f60c97" containerName="kube-state-metrics" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.334761 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="be472357-f01e-4e83-8903-49c0e2f60c97" containerName="kube-state-metrics" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.335148 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="be472357-f01e-4e83-8903-49c0e2f60c97" containerName="kube-state-metrics" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.335913 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.343562 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.343652 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.345750 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.430381 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9h8b\" (UniqueName: \"kubernetes.io/projected/b3edf21b-40ad-40e1-98db-78f609a9b4bc-kube-api-access-b9h8b\") pod \"kube-state-metrics-0\" (UID: \"b3edf21b-40ad-40e1-98db-78f609a9b4bc\") " pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.430468 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3edf21b-40ad-40e1-98db-78f609a9b4bc-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b3edf21b-40ad-40e1-98db-78f609a9b4bc\") " pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.430525 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3edf21b-40ad-40e1-98db-78f609a9b4bc-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b3edf21b-40ad-40e1-98db-78f609a9b4bc\") " pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.430559 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b3edf21b-40ad-40e1-98db-78f609a9b4bc-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b3edf21b-40ad-40e1-98db-78f609a9b4bc\") " pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.531853 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9h8b\" (UniqueName: \"kubernetes.io/projected/b3edf21b-40ad-40e1-98db-78f609a9b4bc-kube-api-access-b9h8b\") pod \"kube-state-metrics-0\" (UID: \"b3edf21b-40ad-40e1-98db-78f609a9b4bc\") " pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.531946 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3edf21b-40ad-40e1-98db-78f609a9b4bc-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b3edf21b-40ad-40e1-98db-78f609a9b4bc\") " pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.531993 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3edf21b-40ad-40e1-98db-78f609a9b4bc-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b3edf21b-40ad-40e1-98db-78f609a9b4bc\") " pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.532034 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b3edf21b-40ad-40e1-98db-78f609a9b4bc-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b3edf21b-40ad-40e1-98db-78f609a9b4bc\") " pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.536999 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b3edf21b-40ad-40e1-98db-78f609a9b4bc-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b3edf21b-40ad-40e1-98db-78f609a9b4bc\") " pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.537058 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3edf21b-40ad-40e1-98db-78f609a9b4bc-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b3edf21b-40ad-40e1-98db-78f609a9b4bc\") " pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.537443 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3edf21b-40ad-40e1-98db-78f609a9b4bc-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b3edf21b-40ad-40e1-98db-78f609a9b4bc\") " pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.550958 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9h8b\" (UniqueName: \"kubernetes.io/projected/b3edf21b-40ad-40e1-98db-78f609a9b4bc-kube-api-access-b9h8b\") pod \"kube-state-metrics-0\" (UID: \"b3edf21b-40ad-40e1-98db-78f609a9b4bc\") " pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.668443 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.796261 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.797193 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="ceilometer-central-agent" containerID="cri-o://02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d" gracePeriod=30 Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.797408 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="proxy-httpd" containerID="cri-o://d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb" gracePeriod=30 Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.797639 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="ceilometer-notification-agent" containerID="cri-o://bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66" gracePeriod=30 Nov 29 00:21:04 crc kubenswrapper[4931]: I1129 00:21:04.797752 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="sg-core" containerID="cri-o://fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50" gracePeriod=30 Nov 29 00:21:05 crc kubenswrapper[4931]: I1129 00:21:05.130310 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 29 00:21:05 crc kubenswrapper[4931]: I1129 00:21:05.221559 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be472357-f01e-4e83-8903-49c0e2f60c97" path="/var/lib/kubelet/pods/be472357-f01e-4e83-8903-49c0e2f60c97/volumes" Nov 29 00:21:05 crc kubenswrapper[4931]: I1129 00:21:05.281845 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b3edf21b-40ad-40e1-98db-78f609a9b4bc","Type":"ContainerStarted","Data":"56f4de6cde7963134bed9bd9164c9dcb76ac7dcaa4bc2dc63593df754e33ca5f"} Nov 29 00:21:05 crc kubenswrapper[4931]: I1129 00:21:05.327252 4931 generic.go:334] "Generic (PLEG): container finished" podID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerID="d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb" exitCode=0 Nov 29 00:21:05 crc kubenswrapper[4931]: I1129 00:21:05.327282 4931 generic.go:334] "Generic (PLEG): container finished" podID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerID="fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50" exitCode=2 Nov 29 00:21:05 crc kubenswrapper[4931]: I1129 00:21:05.327292 4931 generic.go:334] "Generic (PLEG): container finished" podID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerID="02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d" exitCode=0 Nov 29 00:21:05 crc kubenswrapper[4931]: I1129 00:21:05.328069 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f","Type":"ContainerDied","Data":"d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb"} Nov 29 00:21:05 crc kubenswrapper[4931]: I1129 00:21:05.328099 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f","Type":"ContainerDied","Data":"fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50"} Nov 29 00:21:05 crc kubenswrapper[4931]: I1129 00:21:05.328110 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f","Type":"ContainerDied","Data":"02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d"} Nov 29 00:21:05 crc kubenswrapper[4931]: I1129 00:21:05.977524 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.162011 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-combined-ca-bundle\") pod \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.162071 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-config-data\") pod \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.162095 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-sg-core-conf-yaml\") pod \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.162178 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-run-httpd\") pod \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.162209 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-scripts\") pod \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.162225 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-log-httpd\") pod \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.162280 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4p7p\" (UniqueName: \"kubernetes.io/projected/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-kube-api-access-z4p7p\") pod \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\" (UID: \"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f\") " Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.162842 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" (UID: "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.163137 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" (UID: "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.167575 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-scripts" (OuterVolumeSpecName: "scripts") pod "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" (UID: "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.167609 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-kube-api-access-z4p7p" (OuterVolumeSpecName: "kube-api-access-z4p7p") pod "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" (UID: "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f"). InnerVolumeSpecName "kube-api-access-z4p7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.189424 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" (UID: "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.246444 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" (UID: "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.266039 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.266255 4931 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.266314 4931 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.266401 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.266456 4931 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.266510 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4p7p\" (UniqueName: \"kubernetes.io/projected/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-kube-api-access-z4p7p\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.280636 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-config-data" (OuterVolumeSpecName: "config-data") pod "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" (UID: "e52b8c44-4758-44ec-ac4d-2f0e5738fd2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.340445 4931 generic.go:334] "Generic (PLEG): container finished" podID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerID="bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66" exitCode=0 Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.340529 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f","Type":"ContainerDied","Data":"bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66"} Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.340539 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.340577 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e52b8c44-4758-44ec-ac4d-2f0e5738fd2f","Type":"ContainerDied","Data":"a7358132857055de9332bc8e8d72bb285454ca4ef11c7b705006b4816ed01200"} Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.340595 4931 scope.go:117] "RemoveContainer" containerID="d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.345388 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b3edf21b-40ad-40e1-98db-78f609a9b4bc","Type":"ContainerStarted","Data":"21686c4457edb0896a583664b7aad8970a819cc7e327afaf9bee00716a4dd139"} Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.345579 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.368200 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.371988 4931 scope.go:117] "RemoveContainer" containerID="fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.390434 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.051960426 podStartE2EDuration="2.390406778s" podCreationTimestamp="2025-11-29 00:21:04 +0000 UTC" firstStartedPulling="2025-11-29 00:21:05.143460213 +0000 UTC m=+1278.305353435" lastFinishedPulling="2025-11-29 00:21:05.481906555 +0000 UTC m=+1278.643799787" observedRunningTime="2025-11-29 00:21:06.366089044 +0000 UTC m=+1279.527982286" watchObservedRunningTime="2025-11-29 00:21:06.390406778 +0000 UTC m=+1279.552300060" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.397073 4931 scope.go:117] "RemoveContainer" containerID="bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.410548 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.419057 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.421276 4931 scope.go:117] "RemoveContainer" containerID="02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.424904 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:21:06 crc kubenswrapper[4931]: E1129 00:21:06.425539 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="sg-core" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.425559 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="sg-core" Nov 29 00:21:06 crc kubenswrapper[4931]: E1129 00:21:06.425573 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="ceilometer-notification-agent" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.425580 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="ceilometer-notification-agent" Nov 29 00:21:06 crc kubenswrapper[4931]: E1129 00:21:06.425594 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="proxy-httpd" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.425600 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="proxy-httpd" Nov 29 00:21:06 crc kubenswrapper[4931]: E1129 00:21:06.425617 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="ceilometer-central-agent" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.425622 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="ceilometer-central-agent" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.425791 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="proxy-httpd" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.425823 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="ceilometer-central-agent" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.425836 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="ceilometer-notification-agent" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.425845 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" containerName="sg-core" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.427342 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.430724 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.430968 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.431220 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.435727 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.495545 4931 scope.go:117] "RemoveContainer" containerID="d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb" Nov 29 00:21:06 crc kubenswrapper[4931]: E1129 00:21:06.495902 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb\": container with ID starting with d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb not found: ID does not exist" containerID="d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.495927 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb"} err="failed to get container status \"d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb\": rpc error: code = NotFound desc = could not find container \"d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb\": container with ID starting with d138e35f93fdb0cd577eb0e46cd339f7c4bb79e45b3aa9aa291d48ce928677fb not found: ID does not exist" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.495948 4931 scope.go:117] "RemoveContainer" containerID="fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50" Nov 29 00:21:06 crc kubenswrapper[4931]: E1129 00:21:06.496116 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50\": container with ID starting with fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50 not found: ID does not exist" containerID="fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.496136 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50"} err="failed to get container status \"fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50\": rpc error: code = NotFound desc = could not find container \"fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50\": container with ID starting with fc7798e492cf1606f1665b1086986ed1999b0aa0b55323437c1cfa22d93b0b50 not found: ID does not exist" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.496147 4931 scope.go:117] "RemoveContainer" containerID="bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66" Nov 29 00:21:06 crc kubenswrapper[4931]: E1129 00:21:06.496310 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66\": container with ID starting with bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66 not found: ID does not exist" containerID="bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.496327 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66"} err="failed to get container status \"bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66\": rpc error: code = NotFound desc = could not find container \"bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66\": container with ID starting with bf5340e74fca9e6ef27e9dfe8358e95ea5a31d01598b4904cb7ace3f06bdda66 not found: ID does not exist" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.496340 4931 scope.go:117] "RemoveContainer" containerID="02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d" Nov 29 00:21:06 crc kubenswrapper[4931]: E1129 00:21:06.496484 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d\": container with ID starting with 02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d not found: ID does not exist" containerID="02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.496503 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d"} err="failed to get container status \"02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d\": rpc error: code = NotFound desc = could not find container \"02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d\": container with ID starting with 02e29ba03457906edcad3860e023f05c7bfd15983460a4355c128916fa84e58d not found: ID does not exist" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.571391 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.571456 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.571474 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7f18517-445a-4295-90f0-7333785879f2-log-httpd\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.571500 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7f18517-445a-4295-90f0-7333785879f2-run-httpd\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.571538 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.571597 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-scripts\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.571617 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-config-data\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.571636 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjt6g\" (UniqueName: \"kubernetes.io/projected/a7f18517-445a-4295-90f0-7333785879f2-kube-api-access-bjt6g\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.673218 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-scripts\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.674030 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-config-data\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.674061 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjt6g\" (UniqueName: \"kubernetes.io/projected/a7f18517-445a-4295-90f0-7333785879f2-kube-api-access-bjt6g\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.674656 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.674750 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.674778 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7f18517-445a-4295-90f0-7333785879f2-log-httpd\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.674852 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7f18517-445a-4295-90f0-7333785879f2-run-httpd\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.674905 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.675589 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7f18517-445a-4295-90f0-7333785879f2-log-httpd\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.675666 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7f18517-445a-4295-90f0-7333785879f2-run-httpd\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.677907 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-scripts\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.678141 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-config-data\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.679462 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.687871 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.689609 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.691976 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjt6g\" (UniqueName: \"kubernetes.io/projected/a7f18517-445a-4295-90f0-7333785879f2-kube-api-access-bjt6g\") pod \"ceilometer-0\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " pod="openstack/ceilometer-0" Nov 29 00:21:06 crc kubenswrapper[4931]: I1129 00:21:06.862464 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:21:07 crc kubenswrapper[4931]: I1129 00:21:07.223704 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e52b8c44-4758-44ec-ac4d-2f0e5738fd2f" path="/var/lib/kubelet/pods/e52b8c44-4758-44ec-ac4d-2f0e5738fd2f/volumes" Nov 29 00:21:07 crc kubenswrapper[4931]: I1129 00:21:07.378654 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:21:07 crc kubenswrapper[4931]: W1129 00:21:07.386670 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7f18517_445a_4295_90f0_7333785879f2.slice/crio-d22701b6c431eeadbd94fb4cad4f4ce1aec9173e32dcc35f358e9c496d36a4e5 WatchSource:0}: Error finding container d22701b6c431eeadbd94fb4cad4f4ce1aec9173e32dcc35f358e9c496d36a4e5: Status 404 returned error can't find the container with id d22701b6c431eeadbd94fb4cad4f4ce1aec9173e32dcc35f358e9c496d36a4e5 Nov 29 00:21:08 crc kubenswrapper[4931]: I1129 00:21:08.099224 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:21:08 crc kubenswrapper[4931]: E1129 00:21:08.099397 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:21:08 crc kubenswrapper[4931]: E1129 00:21:08.099633 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-764cb44577-7q67f: configmap "swift-ring-files" not found Nov 29 00:21:08 crc kubenswrapper[4931]: E1129 00:21:08.099695 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift podName:550e16f7-f101-4a91-8e2f-45af3430d6b9 nodeName:}" failed. No retries permitted until 2025-11-29 00:22:12.099677982 +0000 UTC m=+1345.261571214 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift") pod "swift-proxy-764cb44577-7q67f" (UID: "550e16f7-f101-4a91-8e2f-45af3430d6b9") : configmap "swift-ring-files" not found Nov 29 00:21:08 crc kubenswrapper[4931]: I1129 00:21:08.364240 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7f18517-445a-4295-90f0-7333785879f2","Type":"ContainerStarted","Data":"1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557"} Nov 29 00:21:08 crc kubenswrapper[4931]: I1129 00:21:08.364308 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7f18517-445a-4295-90f0-7333785879f2","Type":"ContainerStarted","Data":"d22701b6c431eeadbd94fb4cad4f4ce1aec9173e32dcc35f358e9c496d36a4e5"} Nov 29 00:21:09 crc kubenswrapper[4931]: I1129 00:21:09.377372 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7f18517-445a-4295-90f0-7333785879f2","Type":"ContainerStarted","Data":"22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886"} Nov 29 00:21:10 crc kubenswrapper[4931]: I1129 00:21:10.389268 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7f18517-445a-4295-90f0-7333785879f2","Type":"ContainerStarted","Data":"c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee"} Nov 29 00:21:13 crc kubenswrapper[4931]: I1129 00:21:13.747391 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.244160 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-gf7vk"] Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.246512 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.250374 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.250577 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.266665 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-gf7vk"] Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.425854 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-scripts\") pod \"nova-cell0-cell-mapping-gf7vk\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.426317 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82j5j\" (UniqueName: \"kubernetes.io/projected/578b8169-9015-4faa-833c-25d8f7d75d10-kube-api-access-82j5j\") pod \"nova-cell0-cell-mapping-gf7vk\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.426454 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gf7vk\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.426580 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-config-data\") pod \"nova-cell0-cell-mapping-gf7vk\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.429022 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.435191 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.441326 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.456119 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.468960 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7f18517-445a-4295-90f0-7333785879f2","Type":"ContainerStarted","Data":"7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62"} Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.470276 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.479238 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.482334 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.488215 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.521058 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.530314 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gf7vk\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.530427 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ae7271-91fe-4390-a8ff-b39949d8675a-logs\") pod \"nova-api-0\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.530471 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53ae7271-91fe-4390-a8ff-b39949d8675a-config-data\") pod \"nova-api-0\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.530503 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-config-data\") pod \"nova-cell0-cell-mapping-gf7vk\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.530552 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glccm\" (UniqueName: \"kubernetes.io/projected/53ae7271-91fe-4390-a8ff-b39949d8675a-kube-api-access-glccm\") pod \"nova-api-0\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.530612 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ae7271-91fe-4390-a8ff-b39949d8675a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.530688 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-scripts\") pod \"nova-cell0-cell-mapping-gf7vk\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.530736 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82j5j\" (UniqueName: \"kubernetes.io/projected/578b8169-9015-4faa-833c-25d8f7d75d10-kube-api-access-82j5j\") pod \"nova-cell0-cell-mapping-gf7vk\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.533183 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.682120662 podStartE2EDuration="8.533165053s" podCreationTimestamp="2025-11-29 00:21:06 +0000 UTC" firstStartedPulling="2025-11-29 00:21:07.388490595 +0000 UTC m=+1280.550383847" lastFinishedPulling="2025-11-29 00:21:13.239535006 +0000 UTC m=+1286.401428238" observedRunningTime="2025-11-29 00:21:14.507484399 +0000 UTC m=+1287.669377651" watchObservedRunningTime="2025-11-29 00:21:14.533165053 +0000 UTC m=+1287.695058285" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.544971 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-config-data\") pod \"nova-cell0-cell-mapping-gf7vk\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.546465 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-scripts\") pod \"nova-cell0-cell-mapping-gf7vk\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.547766 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gf7vk\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.560738 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82j5j\" (UniqueName: \"kubernetes.io/projected/578b8169-9015-4faa-833c-25d8f7d75d10-kube-api-access-82j5j\") pod \"nova-cell0-cell-mapping-gf7vk\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.565606 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.641231 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glccm\" (UniqueName: \"kubernetes.io/projected/53ae7271-91fe-4390-a8ff-b39949d8675a-kube-api-access-glccm\") pod \"nova-api-0\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.641292 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62l2l\" (UniqueName: \"kubernetes.io/projected/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-kube-api-access-62l2l\") pod \"nova-scheduler-0\" (UID: \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.641327 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.663994 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ae7271-91fe-4390-a8ff-b39949d8675a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.664136 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-config-data\") pod \"nova-scheduler-0\" (UID: \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.664448 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ae7271-91fe-4390-a8ff-b39949d8675a-logs\") pod \"nova-api-0\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.664539 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53ae7271-91fe-4390-a8ff-b39949d8675a-config-data\") pod \"nova-api-0\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.707921 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ae7271-91fe-4390-a8ff-b39949d8675a-logs\") pod \"nova-api-0\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.743063 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.750697 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.758768 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ae7271-91fe-4390-a8ff-b39949d8675a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.758932 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.760036 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53ae7271-91fe-4390-a8ff-b39949d8675a-config-data\") pod \"nova-api-0\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.760551 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glccm\" (UniqueName: \"kubernetes.io/projected/53ae7271-91fe-4390-a8ff-b39949d8675a-kube-api-access-glccm\") pod \"nova-api-0\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " pod="openstack/nova-api-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.774032 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.779003 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62l2l\" (UniqueName: \"kubernetes.io/projected/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-kube-api-access-62l2l\") pod \"nova-scheduler-0\" (UID: \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.779103 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.779225 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-config-data\") pod \"nova-scheduler-0\" (UID: \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.786290 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-config-data\") pod \"nova-scheduler-0\" (UID: \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.791233 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.812385 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62l2l\" (UniqueName: \"kubernetes.io/projected/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-kube-api-access-62l2l\") pod \"nova-scheduler-0\" (UID: \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.836048 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.847462 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.849064 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.859349 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.860381 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.880666 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjsnc\" (UniqueName: \"kubernetes.io/projected/b7862c78-08c4-4503-ae0b-defd34e755c6-kube-api-access-kjsnc\") pod \"nova-metadata-0\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " pod="openstack/nova-metadata-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.881259 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03c88d33-e499-4955-8976-b20c2bd4a09a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"03c88d33-e499-4955-8976-b20c2bd4a09a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.881748 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7862c78-08c4-4503-ae0b-defd34e755c6-config-data\") pod \"nova-metadata-0\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " pod="openstack/nova-metadata-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.881969 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03c88d33-e499-4955-8976-b20c2bd4a09a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"03c88d33-e499-4955-8976-b20c2bd4a09a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.882128 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvj4p\" (UniqueName: \"kubernetes.io/projected/03c88d33-e499-4955-8976-b20c2bd4a09a-kube-api-access-qvj4p\") pod \"nova-cell1-novncproxy-0\" (UID: \"03c88d33-e499-4955-8976-b20c2bd4a09a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.882195 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7862c78-08c4-4503-ae0b-defd34e755c6-logs\") pod \"nova-metadata-0\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " pod="openstack/nova-metadata-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.882289 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7862c78-08c4-4503-ae0b-defd34e755c6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " pod="openstack/nova-metadata-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.901576 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c6ccb6797-cgp75"] Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.903654 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.938643 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6ccb6797-cgp75"] Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.983900 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjsnc\" (UniqueName: \"kubernetes.io/projected/b7862c78-08c4-4503-ae0b-defd34e755c6-kube-api-access-kjsnc\") pod \"nova-metadata-0\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " pod="openstack/nova-metadata-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.983961 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03c88d33-e499-4955-8976-b20c2bd4a09a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"03c88d33-e499-4955-8976-b20c2bd4a09a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.983984 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7862c78-08c4-4503-ae0b-defd34e755c6-config-data\") pod \"nova-metadata-0\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " pod="openstack/nova-metadata-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.984048 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03c88d33-e499-4955-8976-b20c2bd4a09a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"03c88d33-e499-4955-8976-b20c2bd4a09a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.984113 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvj4p\" (UniqueName: \"kubernetes.io/projected/03c88d33-e499-4955-8976-b20c2bd4a09a-kube-api-access-qvj4p\") pod \"nova-cell1-novncproxy-0\" (UID: \"03c88d33-e499-4955-8976-b20c2bd4a09a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.984130 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7862c78-08c4-4503-ae0b-defd34e755c6-logs\") pod \"nova-metadata-0\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " pod="openstack/nova-metadata-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.984158 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7862c78-08c4-4503-ae0b-defd34e755c6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " pod="openstack/nova-metadata-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.987908 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7862c78-08c4-4503-ae0b-defd34e755c6-logs\") pod \"nova-metadata-0\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " pod="openstack/nova-metadata-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.989841 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03c88d33-e499-4955-8976-b20c2bd4a09a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"03c88d33-e499-4955-8976-b20c2bd4a09a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.990303 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7862c78-08c4-4503-ae0b-defd34e755c6-config-data\") pod \"nova-metadata-0\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " pod="openstack/nova-metadata-0" Nov 29 00:21:14 crc kubenswrapper[4931]: I1129 00:21:14.990462 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7862c78-08c4-4503-ae0b-defd34e755c6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " pod="openstack/nova-metadata-0" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.002184 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03c88d33-e499-4955-8976-b20c2bd4a09a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"03c88d33-e499-4955-8976-b20c2bd4a09a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.005721 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvj4p\" (UniqueName: \"kubernetes.io/projected/03c88d33-e499-4955-8976-b20c2bd4a09a-kube-api-access-qvj4p\") pod \"nova-cell1-novncproxy-0\" (UID: \"03c88d33-e499-4955-8976-b20c2bd4a09a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.007193 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjsnc\" (UniqueName: \"kubernetes.io/projected/b7862c78-08c4-4503-ae0b-defd34e755c6-kube-api-access-kjsnc\") pod \"nova-metadata-0\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " pod="openstack/nova-metadata-0" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.059207 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.094772 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4kxn\" (UniqueName: \"kubernetes.io/projected/149e0e4c-6506-46d0-a6a4-965d8966c71e-kube-api-access-l4kxn\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.094835 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-ovsdbserver-sb\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.094865 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-config\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.094944 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-ovsdbserver-nb\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.095079 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-dns-svc\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.117198 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.169254 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.189374 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.196640 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-ovsdbserver-nb\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.196758 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-dns-svc\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.196858 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4kxn\" (UniqueName: \"kubernetes.io/projected/149e0e4c-6506-46d0-a6a4-965d8966c71e-kube-api-access-l4kxn\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.196878 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-ovsdbserver-sb\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.196899 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-config\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.197774 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-config\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.199001 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-dns-svc\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.199157 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-ovsdbserver-sb\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.210626 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-ovsdbserver-nb\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.246285 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4kxn\" (UniqueName: \"kubernetes.io/projected/149e0e4c-6506-46d0-a6a4-965d8966c71e-kube-api-access-l4kxn\") pod \"dnsmasq-dns-7c6ccb6797-cgp75\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.336417 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-gf7vk"] Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.520052 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gf7vk" event={"ID":"578b8169-9015-4faa-833c-25d8f7d75d10","Type":"ContainerStarted","Data":"668997d63a5f74595412e520d5d2285393fbc735c10f7bfedbabc24e889c86ea"} Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.549595 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.711531 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.735310 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rd7sl"] Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.736786 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.739227 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.739279 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.749445 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rd7sl"] Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.927717 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-config-data\") pod \"nova-cell1-conductor-db-sync-rd7sl\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.927767 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvx48\" (UniqueName: \"kubernetes.io/projected/8ce314fe-040a-4d82-925d-d48c52a993dd-kube-api-access-gvx48\") pod \"nova-cell1-conductor-db-sync-rd7sl\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.927821 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-scripts\") pod \"nova-cell1-conductor-db-sync-rd7sl\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.927843 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rd7sl\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:15 crc kubenswrapper[4931]: I1129 00:21:15.948750 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.029884 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-config-data\") pod \"nova-cell1-conductor-db-sync-rd7sl\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.029954 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvx48\" (UniqueName: \"kubernetes.io/projected/8ce314fe-040a-4d82-925d-d48c52a993dd-kube-api-access-gvx48\") pod \"nova-cell1-conductor-db-sync-rd7sl\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.029999 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-scripts\") pod \"nova-cell1-conductor-db-sync-rd7sl\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.030025 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rd7sl\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.040696 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-config-data\") pod \"nova-cell1-conductor-db-sync-rd7sl\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.061538 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.079469 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rd7sl\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.106754 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-scripts\") pod \"nova-cell1-conductor-db-sync-rd7sl\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.109436 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvx48\" (UniqueName: \"kubernetes.io/projected/8ce314fe-040a-4d82-925d-d48c52a993dd-kube-api-access-gvx48\") pod \"nova-cell1-conductor-db-sync-rd7sl\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.112069 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.173663 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6ccb6797-cgp75"] Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.359467 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.601519 4931 generic.go:334] "Generic (PLEG): container finished" podID="149e0e4c-6506-46d0-a6a4-965d8966c71e" containerID="805fc2096b60abcd283157465e921f32416a8cbbb573923e33db514a02922623" exitCode=0 Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.601595 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" event={"ID":"149e0e4c-6506-46d0-a6a4-965d8966c71e","Type":"ContainerDied","Data":"805fc2096b60abcd283157465e921f32416a8cbbb573923e33db514a02922623"} Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.602228 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" event={"ID":"149e0e4c-6506-46d0-a6a4-965d8966c71e","Type":"ContainerStarted","Data":"c599ca4df24ed604688409fdb767083f346f6c4063d4fe3fab6448f726a8bcad"} Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.614059 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"03c88d33-e499-4955-8976-b20c2bd4a09a","Type":"ContainerStarted","Data":"a72161fbce2a015f0fbc6613500e29ba6242f665fbd0c360d16711a25606ba80"} Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.638231 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gf7vk" event={"ID":"578b8169-9015-4faa-833c-25d8f7d75d10","Type":"ContainerStarted","Data":"c687e3ab7c36b06a8733eb3998199211c2735d22306e26ffb70a209f8a330428"} Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.675662 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-gf7vk" podStartSLOduration=2.675637054 podStartE2EDuration="2.675637054s" podCreationTimestamp="2025-11-29 00:21:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:21:16.666955663 +0000 UTC m=+1289.828848915" watchObservedRunningTime="2025-11-29 00:21:16.675637054 +0000 UTC m=+1289.837530286" Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.678361 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"74945ff2-dbac-4c6a-9a6c-616afb19bfe4","Type":"ContainerStarted","Data":"509dc548c412565938c60a774ad5a4b7d99a1281bdaa7b59bd31be110757d3b4"} Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.684340 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7862c78-08c4-4503-ae0b-defd34e755c6","Type":"ContainerStarted","Data":"e3f6614f756869527a511a59c2c0f697b7dfc3bacefd70342e3c7128ad1e6e81"} Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.707146 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"53ae7271-91fe-4390-a8ff-b39949d8675a","Type":"ContainerStarted","Data":"d10e6f41777c1e9d3e95dcd37b6ae19a2ec0a10713f7259aa34609ae5d243135"} Nov 29 00:21:16 crc kubenswrapper[4931]: I1129 00:21:16.903790 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rd7sl"] Nov 29 00:21:16 crc kubenswrapper[4931]: W1129 00:21:16.917174 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ce314fe_040a_4d82_925d_d48c52a993dd.slice/crio-51e8a3b31a1a39a5be6b71f3e1e41a2d08ed8aec60089d74c59ec2f266531f19 WatchSource:0}: Error finding container 51e8a3b31a1a39a5be6b71f3e1e41a2d08ed8aec60089d74c59ec2f266531f19: Status 404 returned error can't find the container with id 51e8a3b31a1a39a5be6b71f3e1e41a2d08ed8aec60089d74c59ec2f266531f19 Nov 29 00:21:17 crc kubenswrapper[4931]: I1129 00:21:17.715634 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rd7sl" event={"ID":"8ce314fe-040a-4d82-925d-d48c52a993dd","Type":"ContainerStarted","Data":"c2ad958301b735a0e4f56dfed8e1511d8d5d02af42a97b9c58bf0ef334a1ff1e"} Nov 29 00:21:17 crc kubenswrapper[4931]: I1129 00:21:17.715977 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rd7sl" event={"ID":"8ce314fe-040a-4d82-925d-d48c52a993dd","Type":"ContainerStarted","Data":"51e8a3b31a1a39a5be6b71f3e1e41a2d08ed8aec60089d74c59ec2f266531f19"} Nov 29 00:21:17 crc kubenswrapper[4931]: I1129 00:21:17.729145 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" event={"ID":"149e0e4c-6506-46d0-a6a4-965d8966c71e","Type":"ContainerStarted","Data":"a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8"} Nov 29 00:21:17 crc kubenswrapper[4931]: I1129 00:21:17.730305 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:17 crc kubenswrapper[4931]: I1129 00:21:17.750480 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-rd7sl" podStartSLOduration=2.750463344 podStartE2EDuration="2.750463344s" podCreationTimestamp="2025-11-29 00:21:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:21:17.747287642 +0000 UTC m=+1290.909180894" watchObservedRunningTime="2025-11-29 00:21:17.750463344 +0000 UTC m=+1290.912356576" Nov 29 00:21:17 crc kubenswrapper[4931]: I1129 00:21:17.771911 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" podStartSLOduration=3.771895375 podStartE2EDuration="3.771895375s" podCreationTimestamp="2025-11-29 00:21:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:21:17.766928301 +0000 UTC m=+1290.928821553" watchObservedRunningTime="2025-11-29 00:21:17.771895375 +0000 UTC m=+1290.933788607" Nov 29 00:21:18 crc kubenswrapper[4931]: I1129 00:21:18.458147 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 00:21:18 crc kubenswrapper[4931]: I1129 00:21:18.480596 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.769840 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"74945ff2-dbac-4c6a-9a6c-616afb19bfe4","Type":"ContainerStarted","Data":"23041759fad0211fd1e958f361032fc88a0cab554d98251ac5139a970438a138"} Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.772979 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7862c78-08c4-4503-ae0b-defd34e755c6","Type":"ContainerStarted","Data":"cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399"} Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.773047 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7862c78-08c4-4503-ae0b-defd34e755c6","Type":"ContainerStarted","Data":"82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a"} Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.773186 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b7862c78-08c4-4503-ae0b-defd34e755c6" containerName="nova-metadata-log" containerID="cri-o://82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a" gracePeriod=30 Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.773251 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b7862c78-08c4-4503-ae0b-defd34e755c6" containerName="nova-metadata-metadata" containerID="cri-o://cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399" gracePeriod=30 Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.791970 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"53ae7271-91fe-4390-a8ff-b39949d8675a","Type":"ContainerStarted","Data":"da47ddcf5db4bd413ff292ed16e42b6fbdcea7fac0fd2eb7ef57aecf5b1f054e"} Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.792043 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"53ae7271-91fe-4390-a8ff-b39949d8675a","Type":"ContainerStarted","Data":"fe8cf2636fb6ca05950e7503341d657922cff8234e607cd782332052fe55a21b"} Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.793799 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.775094664 podStartE2EDuration="7.793778388s" podCreationTimestamp="2025-11-29 00:21:14 +0000 UTC" firstStartedPulling="2025-11-29 00:21:15.717593857 +0000 UTC m=+1288.879487089" lastFinishedPulling="2025-11-29 00:21:20.736277581 +0000 UTC m=+1293.898170813" observedRunningTime="2025-11-29 00:21:21.791728889 +0000 UTC m=+1294.953622141" watchObservedRunningTime="2025-11-29 00:21:21.793778388 +0000 UTC m=+1294.955671620" Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.814963 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"03c88d33-e499-4955-8976-b20c2bd4a09a","Type":"ContainerStarted","Data":"28691ee7f4e971382326216c344002bbde7ae7ff76122bcb79215379b553dedf"} Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.815117 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="03c88d33-e499-4955-8976-b20c2bd4a09a" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://28691ee7f4e971382326216c344002bbde7ae7ff76122bcb79215379b553dedf" gracePeriod=30 Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.827723 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.222073219 podStartE2EDuration="7.82769788s" podCreationTimestamp="2025-11-29 00:21:14 +0000 UTC" firstStartedPulling="2025-11-29 00:21:16.133151252 +0000 UTC m=+1289.295044484" lastFinishedPulling="2025-11-29 00:21:20.738775923 +0000 UTC m=+1293.900669145" observedRunningTime="2025-11-29 00:21:21.814253321 +0000 UTC m=+1294.976146553" watchObservedRunningTime="2025-11-29 00:21:21.82769788 +0000 UTC m=+1294.989591112" Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.869145 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.112213287 podStartE2EDuration="7.86911247s" podCreationTimestamp="2025-11-29 00:21:14 +0000 UTC" firstStartedPulling="2025-11-29 00:21:15.9801276 +0000 UTC m=+1289.142020832" lastFinishedPulling="2025-11-29 00:21:20.737026773 +0000 UTC m=+1293.898920015" observedRunningTime="2025-11-29 00:21:21.834962571 +0000 UTC m=+1294.996855803" watchObservedRunningTime="2025-11-29 00:21:21.86911247 +0000 UTC m=+1295.031005702" Nov 29 00:21:21 crc kubenswrapper[4931]: I1129 00:21:21.920149 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.272098528 podStartE2EDuration="7.919955872s" podCreationTimestamp="2025-11-29 00:21:14 +0000 UTC" firstStartedPulling="2025-11-29 00:21:16.095531493 +0000 UTC m=+1289.257424725" lastFinishedPulling="2025-11-29 00:21:20.743388837 +0000 UTC m=+1293.905282069" observedRunningTime="2025-11-29 00:21:21.855890347 +0000 UTC m=+1295.017783589" watchObservedRunningTime="2025-11-29 00:21:21.919955872 +0000 UTC m=+1295.081849104" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.717487 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.794708 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7862c78-08c4-4503-ae0b-defd34e755c6-config-data\") pod \"b7862c78-08c4-4503-ae0b-defd34e755c6\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.794810 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7862c78-08c4-4503-ae0b-defd34e755c6-combined-ca-bundle\") pod \"b7862c78-08c4-4503-ae0b-defd34e755c6\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.794941 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjsnc\" (UniqueName: \"kubernetes.io/projected/b7862c78-08c4-4503-ae0b-defd34e755c6-kube-api-access-kjsnc\") pod \"b7862c78-08c4-4503-ae0b-defd34e755c6\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.794973 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7862c78-08c4-4503-ae0b-defd34e755c6-logs\") pod \"b7862c78-08c4-4503-ae0b-defd34e755c6\" (UID: \"b7862c78-08c4-4503-ae0b-defd34e755c6\") " Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.796367 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7862c78-08c4-4503-ae0b-defd34e755c6-logs" (OuterVolumeSpecName: "logs") pod "b7862c78-08c4-4503-ae0b-defd34e755c6" (UID: "b7862c78-08c4-4503-ae0b-defd34e755c6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.816128 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7862c78-08c4-4503-ae0b-defd34e755c6-kube-api-access-kjsnc" (OuterVolumeSpecName: "kube-api-access-kjsnc") pod "b7862c78-08c4-4503-ae0b-defd34e755c6" (UID: "b7862c78-08c4-4503-ae0b-defd34e755c6"). InnerVolumeSpecName "kube-api-access-kjsnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.824371 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7862c78-08c4-4503-ae0b-defd34e755c6-config-data" (OuterVolumeSpecName: "config-data") pod "b7862c78-08c4-4503-ae0b-defd34e755c6" (UID: "b7862c78-08c4-4503-ae0b-defd34e755c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.842242 4931 generic.go:334] "Generic (PLEG): container finished" podID="b7862c78-08c4-4503-ae0b-defd34e755c6" containerID="cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399" exitCode=0 Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.842277 4931 generic.go:334] "Generic (PLEG): container finished" podID="b7862c78-08c4-4503-ae0b-defd34e755c6" containerID="82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a" exitCode=143 Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.843092 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7862c78-08c4-4503-ae0b-defd34e755c6","Type":"ContainerDied","Data":"cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399"} Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.843144 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7862c78-08c4-4503-ae0b-defd34e755c6","Type":"ContainerDied","Data":"82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a"} Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.843158 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b7862c78-08c4-4503-ae0b-defd34e755c6","Type":"ContainerDied","Data":"e3f6614f756869527a511a59c2c0f697b7dfc3bacefd70342e3c7128ad1e6e81"} Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.843177 4931 scope.go:117] "RemoveContainer" containerID="cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.843436 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.845562 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7862c78-08c4-4503-ae0b-defd34e755c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7862c78-08c4-4503-ae0b-defd34e755c6" (UID: "b7862c78-08c4-4503-ae0b-defd34e755c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.897799 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7862c78-08c4-4503-ae0b-defd34e755c6-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.898059 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7862c78-08c4-4503-ae0b-defd34e755c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.898072 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjsnc\" (UniqueName: \"kubernetes.io/projected/b7862c78-08c4-4503-ae0b-defd34e755c6-kube-api-access-kjsnc\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.898084 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7862c78-08c4-4503-ae0b-defd34e755c6-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.932946 4931 scope.go:117] "RemoveContainer" containerID="82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.977522 4931 scope.go:117] "RemoveContainer" containerID="cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399" Nov 29 00:21:22 crc kubenswrapper[4931]: E1129 00:21:22.978796 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399\": container with ID starting with cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399 not found: ID does not exist" containerID="cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.978858 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399"} err="failed to get container status \"cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399\": rpc error: code = NotFound desc = could not find container \"cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399\": container with ID starting with cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399 not found: ID does not exist" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.978883 4931 scope.go:117] "RemoveContainer" containerID="82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a" Nov 29 00:21:22 crc kubenswrapper[4931]: E1129 00:21:22.979580 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a\": container with ID starting with 82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a not found: ID does not exist" containerID="82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.979608 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a"} err="failed to get container status \"82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a\": rpc error: code = NotFound desc = could not find container \"82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a\": container with ID starting with 82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a not found: ID does not exist" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.979626 4931 scope.go:117] "RemoveContainer" containerID="cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.980763 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399"} err="failed to get container status \"cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399\": rpc error: code = NotFound desc = could not find container \"cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399\": container with ID starting with cc8f22809e47ae87f099a4c30c0b902255e6dcf4297daa6615a9085114cb9399 not found: ID does not exist" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.980819 4931 scope.go:117] "RemoveContainer" containerID="82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a" Nov 29 00:21:22 crc kubenswrapper[4931]: I1129 00:21:22.981192 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a"} err="failed to get container status \"82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a\": rpc error: code = NotFound desc = could not find container \"82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a\": container with ID starting with 82f156f84b18b4729d991b5a431f213979626ccce0383441e54cdb5e89a5284a not found: ID does not exist" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.178672 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.189311 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.208707 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:23 crc kubenswrapper[4931]: E1129 00:21:23.209121 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7862c78-08c4-4503-ae0b-defd34e755c6" containerName="nova-metadata-metadata" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.209135 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7862c78-08c4-4503-ae0b-defd34e755c6" containerName="nova-metadata-metadata" Nov 29 00:21:23 crc kubenswrapper[4931]: E1129 00:21:23.209147 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7862c78-08c4-4503-ae0b-defd34e755c6" containerName="nova-metadata-log" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.209153 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7862c78-08c4-4503-ae0b-defd34e755c6" containerName="nova-metadata-log" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.209323 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7862c78-08c4-4503-ae0b-defd34e755c6" containerName="nova-metadata-log" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.209355 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7862c78-08c4-4503-ae0b-defd34e755c6" containerName="nova-metadata-metadata" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.210447 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.214350 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.214741 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.224385 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7862c78-08c4-4503-ae0b-defd34e755c6" path="/var/lib/kubelet/pods/b7862c78-08c4-4503-ae0b-defd34e755c6/volumes" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.225112 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.306209 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5442f155-f4c6-4331-a127-5a3c1832d1dc-logs\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.306273 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.306409 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlc7m\" (UniqueName: \"kubernetes.io/projected/5442f155-f4c6-4331-a127-5a3c1832d1dc-kube-api-access-vlc7m\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.306447 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.306476 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-config-data\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.409158 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlc7m\" (UniqueName: \"kubernetes.io/projected/5442f155-f4c6-4331-a127-5a3c1832d1dc-kube-api-access-vlc7m\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.409920 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.410895 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-config-data\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.411262 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5442f155-f4c6-4331-a127-5a3c1832d1dc-logs\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.411328 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.411975 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5442f155-f4c6-4331-a127-5a3c1832d1dc-logs\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.415135 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-config-data\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.416192 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.416931 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.432379 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlc7m\" (UniqueName: \"kubernetes.io/projected/5442f155-f4c6-4331-a127-5a3c1832d1dc-kube-api-access-vlc7m\") pod \"nova-metadata-0\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.535719 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.854434 4931 generic.go:334] "Generic (PLEG): container finished" podID="578b8169-9015-4faa-833c-25d8f7d75d10" containerID="c687e3ab7c36b06a8733eb3998199211c2735d22306e26ffb70a209f8a330428" exitCode=0 Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.854530 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gf7vk" event={"ID":"578b8169-9015-4faa-833c-25d8f7d75d10","Type":"ContainerDied","Data":"c687e3ab7c36b06a8733eb3998199211c2735d22306e26ffb70a209f8a330428"} Nov 29 00:21:23 crc kubenswrapper[4931]: I1129 00:21:23.960322 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:23 crc kubenswrapper[4931]: W1129 00:21:23.968758 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5442f155_f4c6_4331_a127_5a3c1832d1dc.slice/crio-8ab4d8d6c799d24192157e95eef2823e72ce7c40ed53a37c099967be84d2126f WatchSource:0}: Error finding container 8ab4d8d6c799d24192157e95eef2823e72ce7c40ed53a37c099967be84d2126f: Status 404 returned error can't find the container with id 8ab4d8d6c799d24192157e95eef2823e72ce7c40ed53a37c099967be84d2126f Nov 29 00:21:24 crc kubenswrapper[4931]: I1129 00:21:24.868711 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5442f155-f4c6-4331-a127-5a3c1832d1dc","Type":"ContainerStarted","Data":"928440447fae8581dee16eaf9a28be17570e240cbdac22bfa7f79ce828c80dde"} Nov 29 00:21:24 crc kubenswrapper[4931]: I1129 00:21:24.869082 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5442f155-f4c6-4331-a127-5a3c1832d1dc","Type":"ContainerStarted","Data":"e7d2b957ede388e8fe9b73b67afd2a627fc22df45105975921c1c7fe66434928"} Nov 29 00:21:24 crc kubenswrapper[4931]: I1129 00:21:24.869098 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5442f155-f4c6-4331-a127-5a3c1832d1dc","Type":"ContainerStarted","Data":"8ab4d8d6c799d24192157e95eef2823e72ce7c40ed53a37c099967be84d2126f"} Nov 29 00:21:24 crc kubenswrapper[4931]: I1129 00:21:24.896972 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.896948349 podStartE2EDuration="1.896948349s" podCreationTimestamp="2025-11-29 00:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:21:24.895126426 +0000 UTC m=+1298.057019678" watchObservedRunningTime="2025-11-29 00:21:24.896948349 +0000 UTC m=+1298.058841581" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.060316 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.060669 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.118099 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.118149 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.170339 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.176304 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.250395 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.359945 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-scripts\") pod \"578b8169-9015-4faa-833c-25d8f7d75d10\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.360123 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82j5j\" (UniqueName: \"kubernetes.io/projected/578b8169-9015-4faa-833c-25d8f7d75d10-kube-api-access-82j5j\") pod \"578b8169-9015-4faa-833c-25d8f7d75d10\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.360238 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-config-data\") pod \"578b8169-9015-4faa-833c-25d8f7d75d10\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.360268 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-combined-ca-bundle\") pod \"578b8169-9015-4faa-833c-25d8f7d75d10\" (UID: \"578b8169-9015-4faa-833c-25d8f7d75d10\") " Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.368137 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/578b8169-9015-4faa-833c-25d8f7d75d10-kube-api-access-82j5j" (OuterVolumeSpecName: "kube-api-access-82j5j") pod "578b8169-9015-4faa-833c-25d8f7d75d10" (UID: "578b8169-9015-4faa-833c-25d8f7d75d10"). InnerVolumeSpecName "kube-api-access-82j5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.368317 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-scripts" (OuterVolumeSpecName: "scripts") pod "578b8169-9015-4faa-833c-25d8f7d75d10" (UID: "578b8169-9015-4faa-833c-25d8f7d75d10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.394947 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-config-data" (OuterVolumeSpecName: "config-data") pod "578b8169-9015-4faa-833c-25d8f7d75d10" (UID: "578b8169-9015-4faa-833c-25d8f7d75d10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.397944 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "578b8169-9015-4faa-833c-25d8f7d75d10" (UID: "578b8169-9015-4faa-833c-25d8f7d75d10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.474023 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.474056 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82j5j\" (UniqueName: \"kubernetes.io/projected/578b8169-9015-4faa-833c-25d8f7d75d10-kube-api-access-82j5j\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.474070 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.474080 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578b8169-9015-4faa-833c-25d8f7d75d10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.551841 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.610019 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77c9c856fc-g4tr2"] Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.610614 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" podUID="7cac2862-eb11-4736-b9cb-68f5c99fdffb" containerName="dnsmasq-dns" containerID="cri-o://1d222546697e9082847b7137f585111ab5a4f8930ca3736f92eb49013870090f" gracePeriod=10 Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.899373 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gf7vk" event={"ID":"578b8169-9015-4faa-833c-25d8f7d75d10","Type":"ContainerDied","Data":"668997d63a5f74595412e520d5d2285393fbc735c10f7bfedbabc24e889c86ea"} Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.899458 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="668997d63a5f74595412e520d5d2285393fbc735c10f7bfedbabc24e889c86ea" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.899418 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gf7vk" Nov 29 00:21:25 crc kubenswrapper[4931]: I1129 00:21:25.950758 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.046276 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.046525 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="53ae7271-91fe-4390-a8ff-b39949d8675a" containerName="nova-api-log" containerID="cri-o://fe8cf2636fb6ca05950e7503341d657922cff8234e607cd782332052fe55a21b" gracePeriod=30 Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.046618 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="53ae7271-91fe-4390-a8ff-b39949d8675a" containerName="nova-api-api" containerID="cri-o://da47ddcf5db4bd413ff292ed16e42b6fbdcea7fac0fd2eb7ef57aecf5b1f054e" gracePeriod=30 Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.049716 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="53ae7271-91fe-4390-a8ff-b39949d8675a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": EOF" Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.049997 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="53ae7271-91fe-4390-a8ff-b39949d8675a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": EOF" Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.065678 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.399228 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.906790 4931 generic.go:334] "Generic (PLEG): container finished" podID="53ae7271-91fe-4390-a8ff-b39949d8675a" containerID="fe8cf2636fb6ca05950e7503341d657922cff8234e607cd782332052fe55a21b" exitCode=143 Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.906870 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"53ae7271-91fe-4390-a8ff-b39949d8675a","Type":"ContainerDied","Data":"fe8cf2636fb6ca05950e7503341d657922cff8234e607cd782332052fe55a21b"} Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.910925 4931 generic.go:334] "Generic (PLEG): container finished" podID="8ce314fe-040a-4d82-925d-d48c52a993dd" containerID="c2ad958301b735a0e4f56dfed8e1511d8d5d02af42a97b9c58bf0ef334a1ff1e" exitCode=0 Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.911010 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rd7sl" event={"ID":"8ce314fe-040a-4d82-925d-d48c52a993dd","Type":"ContainerDied","Data":"c2ad958301b735a0e4f56dfed8e1511d8d5d02af42a97b9c58bf0ef334a1ff1e"} Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.911129 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5442f155-f4c6-4331-a127-5a3c1832d1dc" containerName="nova-metadata-log" containerID="cri-o://e7d2b957ede388e8fe9b73b67afd2a627fc22df45105975921c1c7fe66434928" gracePeriod=30 Nov 29 00:21:26 crc kubenswrapper[4931]: I1129 00:21:26.911251 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5442f155-f4c6-4331-a127-5a3c1832d1dc" containerName="nova-metadata-metadata" containerID="cri-o://928440447fae8581dee16eaf9a28be17570e240cbdac22bfa7f79ce828c80dde" gracePeriod=30 Nov 29 00:21:27 crc kubenswrapper[4931]: I1129 00:21:27.922247 4931 generic.go:334] "Generic (PLEG): container finished" podID="5442f155-f4c6-4331-a127-5a3c1832d1dc" containerID="928440447fae8581dee16eaf9a28be17570e240cbdac22bfa7f79ce828c80dde" exitCode=0 Nov 29 00:21:27 crc kubenswrapper[4931]: I1129 00:21:27.922531 4931 generic.go:334] "Generic (PLEG): container finished" podID="5442f155-f4c6-4331-a127-5a3c1832d1dc" containerID="e7d2b957ede388e8fe9b73b67afd2a627fc22df45105975921c1c7fe66434928" exitCode=143 Nov 29 00:21:27 crc kubenswrapper[4931]: I1129 00:21:27.922377 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5442f155-f4c6-4331-a127-5a3c1832d1dc","Type":"ContainerDied","Data":"928440447fae8581dee16eaf9a28be17570e240cbdac22bfa7f79ce828c80dde"} Nov 29 00:21:27 crc kubenswrapper[4931]: I1129 00:21:27.922598 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5442f155-f4c6-4331-a127-5a3c1832d1dc","Type":"ContainerDied","Data":"e7d2b957ede388e8fe9b73b67afd2a627fc22df45105975921c1c7fe66434928"} Nov 29 00:21:27 crc kubenswrapper[4931]: I1129 00:21:27.922613 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5442f155-f4c6-4331-a127-5a3c1832d1dc","Type":"ContainerDied","Data":"8ab4d8d6c799d24192157e95eef2823e72ce7c40ed53a37c099967be84d2126f"} Nov 29 00:21:27 crc kubenswrapper[4931]: I1129 00:21:27.922623 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ab4d8d6c799d24192157e95eef2823e72ce7c40ed53a37c099967be84d2126f" Nov 29 00:21:27 crc kubenswrapper[4931]: I1129 00:21:27.927840 4931 generic.go:334] "Generic (PLEG): container finished" podID="7cac2862-eb11-4736-b9cb-68f5c99fdffb" containerID="1d222546697e9082847b7137f585111ab5a4f8930ca3736f92eb49013870090f" exitCode=0 Nov 29 00:21:27 crc kubenswrapper[4931]: I1129 00:21:27.927889 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" event={"ID":"7cac2862-eb11-4736-b9cb-68f5c99fdffb","Type":"ContainerDied","Data":"1d222546697e9082847b7137f585111ab5a4f8930ca3736f92eb49013870090f"} Nov 29 00:21:27 crc kubenswrapper[4931]: I1129 00:21:27.928105 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="74945ff2-dbac-4c6a-9a6c-616afb19bfe4" containerName="nova-scheduler-scheduler" containerID="cri-o://23041759fad0211fd1e958f361032fc88a0cab554d98251ac5139a970438a138" gracePeriod=30 Nov 29 00:21:27 crc kubenswrapper[4931]: I1129 00:21:27.947374 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.014565 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.022358 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-config-data\") pod \"5442f155-f4c6-4331-a127-5a3c1832d1dc\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.022518 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlc7m\" (UniqueName: \"kubernetes.io/projected/5442f155-f4c6-4331-a127-5a3c1832d1dc-kube-api-access-vlc7m\") pod \"5442f155-f4c6-4331-a127-5a3c1832d1dc\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.022554 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-combined-ca-bundle\") pod \"5442f155-f4c6-4331-a127-5a3c1832d1dc\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.022662 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-nova-metadata-tls-certs\") pod \"5442f155-f4c6-4331-a127-5a3c1832d1dc\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.022696 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5442f155-f4c6-4331-a127-5a3c1832d1dc-logs\") pod \"5442f155-f4c6-4331-a127-5a3c1832d1dc\" (UID: \"5442f155-f4c6-4331-a127-5a3c1832d1dc\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.023723 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5442f155-f4c6-4331-a127-5a3c1832d1dc-logs" (OuterVolumeSpecName: "logs") pod "5442f155-f4c6-4331-a127-5a3c1832d1dc" (UID: "5442f155-f4c6-4331-a127-5a3c1832d1dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.037167 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5442f155-f4c6-4331-a127-5a3c1832d1dc-kube-api-access-vlc7m" (OuterVolumeSpecName: "kube-api-access-vlc7m") pod "5442f155-f4c6-4331-a127-5a3c1832d1dc" (UID: "5442f155-f4c6-4331-a127-5a3c1832d1dc"). InnerVolumeSpecName "kube-api-access-vlc7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.054607 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-config-data" (OuterVolumeSpecName: "config-data") pod "5442f155-f4c6-4331-a127-5a3c1832d1dc" (UID: "5442f155-f4c6-4331-a127-5a3c1832d1dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.084990 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5442f155-f4c6-4331-a127-5a3c1832d1dc" (UID: "5442f155-f4c6-4331-a127-5a3c1832d1dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.090532 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "5442f155-f4c6-4331-a127-5a3c1832d1dc" (UID: "5442f155-f4c6-4331-a127-5a3c1832d1dc"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.125803 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-ovsdbserver-nb\") pod \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.125917 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-ovsdbserver-sb\") pod \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.125970 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lndz\" (UniqueName: \"kubernetes.io/projected/7cac2862-eb11-4736-b9cb-68f5c99fdffb-kube-api-access-5lndz\") pod \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.125989 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-config\") pod \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.126051 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-dns-svc\") pod \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\" (UID: \"7cac2862-eb11-4736-b9cb-68f5c99fdffb\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.126632 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlc7m\" (UniqueName: \"kubernetes.io/projected/5442f155-f4c6-4331-a127-5a3c1832d1dc-kube-api-access-vlc7m\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.126654 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.126664 4931 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.126675 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5442f155-f4c6-4331-a127-5a3c1832d1dc-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.126683 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5442f155-f4c6-4331-a127-5a3c1832d1dc-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.134316 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cac2862-eb11-4736-b9cb-68f5c99fdffb-kube-api-access-5lndz" (OuterVolumeSpecName: "kube-api-access-5lndz") pod "7cac2862-eb11-4736-b9cb-68f5c99fdffb" (UID: "7cac2862-eb11-4736-b9cb-68f5c99fdffb"). InnerVolumeSpecName "kube-api-access-5lndz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.174722 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7cac2862-eb11-4736-b9cb-68f5c99fdffb" (UID: "7cac2862-eb11-4736-b9cb-68f5c99fdffb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.185453 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7cac2862-eb11-4736-b9cb-68f5c99fdffb" (UID: "7cac2862-eb11-4736-b9cb-68f5c99fdffb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.189339 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-config" (OuterVolumeSpecName: "config") pod "7cac2862-eb11-4736-b9cb-68f5c99fdffb" (UID: "7cac2862-eb11-4736-b9cb-68f5c99fdffb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.190625 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7cac2862-eb11-4736-b9cb-68f5c99fdffb" (UID: "7cac2862-eb11-4736-b9cb-68f5c99fdffb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.233744 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lndz\" (UniqueName: \"kubernetes.io/projected/7cac2862-eb11-4736-b9cb-68f5c99fdffb-kube-api-access-5lndz\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.233774 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.233786 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.233797 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.233829 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cac2862-eb11-4736-b9cb-68f5c99fdffb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.256709 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.335428 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-combined-ca-bundle\") pod \"8ce314fe-040a-4d82-925d-d48c52a993dd\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.335602 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-scripts\") pod \"8ce314fe-040a-4d82-925d-d48c52a993dd\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.335653 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-config-data\") pod \"8ce314fe-040a-4d82-925d-d48c52a993dd\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.335737 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvx48\" (UniqueName: \"kubernetes.io/projected/8ce314fe-040a-4d82-925d-d48c52a993dd-kube-api-access-gvx48\") pod \"8ce314fe-040a-4d82-925d-d48c52a993dd\" (UID: \"8ce314fe-040a-4d82-925d-d48c52a993dd\") " Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.338656 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-scripts" (OuterVolumeSpecName: "scripts") pod "8ce314fe-040a-4d82-925d-d48c52a993dd" (UID: "8ce314fe-040a-4d82-925d-d48c52a993dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.342594 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ce314fe-040a-4d82-925d-d48c52a993dd-kube-api-access-gvx48" (OuterVolumeSpecName: "kube-api-access-gvx48") pod "8ce314fe-040a-4d82-925d-d48c52a993dd" (UID: "8ce314fe-040a-4d82-925d-d48c52a993dd"). InnerVolumeSpecName "kube-api-access-gvx48". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.361995 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ce314fe-040a-4d82-925d-d48c52a993dd" (UID: "8ce314fe-040a-4d82-925d-d48c52a993dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.364365 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-config-data" (OuterVolumeSpecName: "config-data") pod "8ce314fe-040a-4d82-925d-d48c52a993dd" (UID: "8ce314fe-040a-4d82-925d-d48c52a993dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.438639 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.438674 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.438686 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvx48\" (UniqueName: \"kubernetes.io/projected/8ce314fe-040a-4d82-925d-d48c52a993dd-kube-api-access-gvx48\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.438696 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce314fe-040a-4d82-925d-d48c52a993dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.940722 4931 generic.go:334] "Generic (PLEG): container finished" podID="74945ff2-dbac-4c6a-9a6c-616afb19bfe4" containerID="23041759fad0211fd1e958f361032fc88a0cab554d98251ac5139a970438a138" exitCode=0 Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.940801 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"74945ff2-dbac-4c6a-9a6c-616afb19bfe4","Type":"ContainerDied","Data":"23041759fad0211fd1e958f361032fc88a0cab554d98251ac5139a970438a138"} Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.944832 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" event={"ID":"7cac2862-eb11-4736-b9cb-68f5c99fdffb","Type":"ContainerDied","Data":"0434d11badb46d2e34d22777e0051fab4f20819819b17e710b04c954b090d43c"} Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.944888 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77c9c856fc-g4tr2" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.944897 4931 scope.go:117] "RemoveContainer" containerID="1d222546697e9082847b7137f585111ab5a4f8930ca3736f92eb49013870090f" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.946813 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.947962 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rd7sl" event={"ID":"8ce314fe-040a-4d82-925d-d48c52a993dd","Type":"ContainerDied","Data":"51e8a3b31a1a39a5be6b71f3e1e41a2d08ed8aec60089d74c59ec2f266531f19"} Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.948000 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51e8a3b31a1a39a5be6b71f3e1e41a2d08ed8aec60089d74c59ec2f266531f19" Nov 29 00:21:28 crc kubenswrapper[4931]: I1129 00:21:28.948059 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rd7sl" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.023311 4931 scope.go:117] "RemoveContainer" containerID="25b38f4b06a00bda834703de4f226534cc327513f90975a574b9966dc92c1ae9" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.064315 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 29 00:21:29 crc kubenswrapper[4931]: E1129 00:21:29.065106 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5442f155-f4c6-4331-a127-5a3c1832d1dc" containerName="nova-metadata-metadata" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.065124 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="5442f155-f4c6-4331-a127-5a3c1832d1dc" containerName="nova-metadata-metadata" Nov 29 00:21:29 crc kubenswrapper[4931]: E1129 00:21:29.065154 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="578b8169-9015-4faa-833c-25d8f7d75d10" containerName="nova-manage" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.065163 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="578b8169-9015-4faa-833c-25d8f7d75d10" containerName="nova-manage" Nov 29 00:21:29 crc kubenswrapper[4931]: E1129 00:21:29.065176 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5442f155-f4c6-4331-a127-5a3c1832d1dc" containerName="nova-metadata-log" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.065182 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="5442f155-f4c6-4331-a127-5a3c1832d1dc" containerName="nova-metadata-log" Nov 29 00:21:29 crc kubenswrapper[4931]: E1129 00:21:29.065200 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cac2862-eb11-4736-b9cb-68f5c99fdffb" containerName="dnsmasq-dns" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.065220 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cac2862-eb11-4736-b9cb-68f5c99fdffb" containerName="dnsmasq-dns" Nov 29 00:21:29 crc kubenswrapper[4931]: E1129 00:21:29.065237 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cac2862-eb11-4736-b9cb-68f5c99fdffb" containerName="init" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.065244 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cac2862-eb11-4736-b9cb-68f5c99fdffb" containerName="init" Nov 29 00:21:29 crc kubenswrapper[4931]: E1129 00:21:29.065262 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce314fe-040a-4d82-925d-d48c52a993dd" containerName="nova-cell1-conductor-db-sync" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.065268 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce314fe-040a-4d82-925d-d48c52a993dd" containerName="nova-cell1-conductor-db-sync" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.065507 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cac2862-eb11-4736-b9cb-68f5c99fdffb" containerName="dnsmasq-dns" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.065543 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ce314fe-040a-4d82-925d-d48c52a993dd" containerName="nova-cell1-conductor-db-sync" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.065551 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="5442f155-f4c6-4331-a127-5a3c1832d1dc" containerName="nova-metadata-metadata" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.065561 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="5442f155-f4c6-4331-a127-5a3c1832d1dc" containerName="nova-metadata-log" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.065578 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="578b8169-9015-4faa-833c-25d8f7d75d10" containerName="nova-manage" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.066437 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.069470 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.083573 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.101585 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.109135 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.121062 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.127166 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.129599 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.129927 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.136968 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.154934 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77c9c856fc-g4tr2"] Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.162780 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2a7ac81-47a3-4c55-90bc-f19842c87e37-logs\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.162877 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecb51737-365b-4866-9d46-3e3aa5729b38-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ecb51737-365b-4866-9d46-3e3aa5729b38\") " pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.162950 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9t68\" (UniqueName: \"kubernetes.io/projected/c2a7ac81-47a3-4c55-90bc-f19842c87e37-kube-api-access-r9t68\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.162998 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9k7b\" (UniqueName: \"kubernetes.io/projected/ecb51737-365b-4866-9d46-3e3aa5729b38-kube-api-access-z9k7b\") pod \"nova-cell1-conductor-0\" (UID: \"ecb51737-365b-4866-9d46-3e3aa5729b38\") " pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.163045 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.163084 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.163117 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb51737-365b-4866-9d46-3e3aa5729b38-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ecb51737-365b-4866-9d46-3e3aa5729b38\") " pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.163137 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-config-data\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.193320 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77c9c856fc-g4tr2"] Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.225156 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5442f155-f4c6-4331-a127-5a3c1832d1dc" path="/var/lib/kubelet/pods/5442f155-f4c6-4331-a127-5a3c1832d1dc/volumes" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.225933 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cac2862-eb11-4736-b9cb-68f5c99fdffb" path="/var/lib/kubelet/pods/7cac2862-eb11-4736-b9cb-68f5c99fdffb/volumes" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.227028 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.266102 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-config-data\") pod \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\" (UID: \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\") " Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.266370 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-combined-ca-bundle\") pod \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\" (UID: \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\") " Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.266485 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62l2l\" (UniqueName: \"kubernetes.io/projected/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-kube-api-access-62l2l\") pod \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\" (UID: \"74945ff2-dbac-4c6a-9a6c-616afb19bfe4\") " Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.269866 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9t68\" (UniqueName: \"kubernetes.io/projected/c2a7ac81-47a3-4c55-90bc-f19842c87e37-kube-api-access-r9t68\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.269945 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9k7b\" (UniqueName: \"kubernetes.io/projected/ecb51737-365b-4866-9d46-3e3aa5729b38-kube-api-access-z9k7b\") pod \"nova-cell1-conductor-0\" (UID: \"ecb51737-365b-4866-9d46-3e3aa5729b38\") " pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.270020 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.270093 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.270160 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb51737-365b-4866-9d46-3e3aa5729b38-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ecb51737-365b-4866-9d46-3e3aa5729b38\") " pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.270190 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-config-data\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.270292 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2a7ac81-47a3-4c55-90bc-f19842c87e37-logs\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.270361 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecb51737-365b-4866-9d46-3e3aa5729b38-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ecb51737-365b-4866-9d46-3e3aa5729b38\") " pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.272549 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2a7ac81-47a3-4c55-90bc-f19842c87e37-logs\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.275901 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.277978 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecb51737-365b-4866-9d46-3e3aa5729b38-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ecb51737-365b-4866-9d46-3e3aa5729b38\") " pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.275048 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-kube-api-access-62l2l" (OuterVolumeSpecName: "kube-api-access-62l2l") pod "74945ff2-dbac-4c6a-9a6c-616afb19bfe4" (UID: "74945ff2-dbac-4c6a-9a6c-616afb19bfe4"). InnerVolumeSpecName "kube-api-access-62l2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.280012 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-config-data\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.283553 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.285641 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb51737-365b-4866-9d46-3e3aa5729b38-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ecb51737-365b-4866-9d46-3e3aa5729b38\") " pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.290858 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9k7b\" (UniqueName: \"kubernetes.io/projected/ecb51737-365b-4866-9d46-3e3aa5729b38-kube-api-access-z9k7b\") pod \"nova-cell1-conductor-0\" (UID: \"ecb51737-365b-4866-9d46-3e3aa5729b38\") " pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.293880 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9t68\" (UniqueName: \"kubernetes.io/projected/c2a7ac81-47a3-4c55-90bc-f19842c87e37-kube-api-access-r9t68\") pod \"nova-metadata-0\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.300168 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-config-data" (OuterVolumeSpecName: "config-data") pod "74945ff2-dbac-4c6a-9a6c-616afb19bfe4" (UID: "74945ff2-dbac-4c6a-9a6c-616afb19bfe4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.319383 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74945ff2-dbac-4c6a-9a6c-616afb19bfe4" (UID: "74945ff2-dbac-4c6a-9a6c-616afb19bfe4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.372249 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.372297 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62l2l\" (UniqueName: \"kubernetes.io/projected/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-kube-api-access-62l2l\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.372313 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74945ff2-dbac-4c6a-9a6c-616afb19bfe4-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.427838 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.525625 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.711663 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 29 00:21:29 crc kubenswrapper[4931]: W1129 00:21:29.715785 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecb51737_365b_4866_9d46_3e3aa5729b38.slice/crio-86c9defe2cb0fa4daa7b148676ca9d9c4108f7b5bad69d607e9c5810ce68c359 WatchSource:0}: Error finding container 86c9defe2cb0fa4daa7b148676ca9d9c4108f7b5bad69d607e9c5810ce68c359: Status 404 returned error can't find the container with id 86c9defe2cb0fa4daa7b148676ca9d9c4108f7b5bad69d607e9c5810ce68c359 Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.958207 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"74945ff2-dbac-4c6a-9a6c-616afb19bfe4","Type":"ContainerDied","Data":"509dc548c412565938c60a774ad5a4b7d99a1281bdaa7b59bd31be110757d3b4"} Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.958265 4931 scope.go:117] "RemoveContainer" containerID="23041759fad0211fd1e958f361032fc88a0cab554d98251ac5139a970438a138" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.958268 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.961908 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ecb51737-365b-4866-9d46-3e3aa5729b38","Type":"ContainerStarted","Data":"f02c3415565ee7d833c99ec504b1b1c2c34da5774b5f128800a07883455acaed"} Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.961960 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ecb51737-365b-4866-9d46-3e3aa5729b38","Type":"ContainerStarted","Data":"86c9defe2cb0fa4daa7b148676ca9d9c4108f7b5bad69d607e9c5810ce68c359"} Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.962054 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:29 crc kubenswrapper[4931]: I1129 00:21:29.981550 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.9815312390000002 podStartE2EDuration="1.981531239s" podCreationTimestamp="2025-11-29 00:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:21:29.978366367 +0000 UTC m=+1303.140259609" watchObservedRunningTime="2025-11-29 00:21:29.981531239 +0000 UTC m=+1303.143424471" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.017231 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.034554 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.042357 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:21:30 crc kubenswrapper[4931]: E1129 00:21:30.042799 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74945ff2-dbac-4c6a-9a6c-616afb19bfe4" containerName="nova-scheduler-scheduler" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.042834 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="74945ff2-dbac-4c6a-9a6c-616afb19bfe4" containerName="nova-scheduler-scheduler" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.043026 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="74945ff2-dbac-4c6a-9a6c-616afb19bfe4" containerName="nova-scheduler-scheduler" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.043634 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.045384 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 29 00:21:30 crc kubenswrapper[4931]: W1129 00:21:30.046150 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2a7ac81_47a3_4c55_90bc_f19842c87e37.slice/crio-8fe866567202b637e05870e2338406b163d6f4f6c65a24ad70f257759765f132 WatchSource:0}: Error finding container 8fe866567202b637e05870e2338406b163d6f4f6c65a24ad70f257759765f132: Status 404 returned error can't find the container with id 8fe866567202b637e05870e2338406b163d6f4f6c65a24ad70f257759765f132 Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.052628 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.078035 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.086594 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.088283 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdk7w\" (UniqueName: \"kubernetes.io/projected/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-kube-api-access-gdk7w\") pod \"nova-scheduler-0\" (UID: \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.088358 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-config-data\") pod \"nova-scheduler-0\" (UID: \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.190008 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.190131 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdk7w\" (UniqueName: \"kubernetes.io/projected/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-kube-api-access-gdk7w\") pod \"nova-scheduler-0\" (UID: \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.190174 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-config-data\") pod \"nova-scheduler-0\" (UID: \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.196261 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.197324 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-config-data\") pod \"nova-scheduler-0\" (UID: \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.208553 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdk7w\" (UniqueName: \"kubernetes.io/projected/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-kube-api-access-gdk7w\") pod \"nova-scheduler-0\" (UID: \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\") " pod="openstack/nova-scheduler-0" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.467836 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.930212 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:21:30 crc kubenswrapper[4931]: W1129 00:21:30.937097 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd62bbf08_1341_44a7_b3d2_d85b0490f6f3.slice/crio-759b79ed53cd355b72261053c0ef3918b19147755c3d0b351a346c729c78722f WatchSource:0}: Error finding container 759b79ed53cd355b72261053c0ef3918b19147755c3d0b351a346c729c78722f: Status 404 returned error can't find the container with id 759b79ed53cd355b72261053c0ef3918b19147755c3d0b351a346c729c78722f Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.971986 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d62bbf08-1341-44a7-b3d2-d85b0490f6f3","Type":"ContainerStarted","Data":"759b79ed53cd355b72261053c0ef3918b19147755c3d0b351a346c729c78722f"} Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.974179 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2a7ac81-47a3-4c55-90bc-f19842c87e37","Type":"ContainerStarted","Data":"88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b"} Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.974209 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2a7ac81-47a3-4c55-90bc-f19842c87e37","Type":"ContainerStarted","Data":"65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff"} Nov 29 00:21:30 crc kubenswrapper[4931]: I1129 00:21:30.974223 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2a7ac81-47a3-4c55-90bc-f19842c87e37","Type":"ContainerStarted","Data":"8fe866567202b637e05870e2338406b163d6f4f6c65a24ad70f257759765f132"} Nov 29 00:21:31 crc kubenswrapper[4931]: I1129 00:21:31.006955 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.006934757 podStartE2EDuration="2.006934757s" podCreationTimestamp="2025-11-29 00:21:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:21:30.99116159 +0000 UTC m=+1304.153054822" watchObservedRunningTime="2025-11-29 00:21:31.006934757 +0000 UTC m=+1304.168827999" Nov 29 00:21:31 crc kubenswrapper[4931]: I1129 00:21:31.230030 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74945ff2-dbac-4c6a-9a6c-616afb19bfe4" path="/var/lib/kubelet/pods/74945ff2-dbac-4c6a-9a6c-616afb19bfe4/volumes" Nov 29 00:21:31 crc kubenswrapper[4931]: I1129 00:21:31.982589 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d62bbf08-1341-44a7-b3d2-d85b0490f6f3","Type":"ContainerStarted","Data":"01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac"} Nov 29 00:21:31 crc kubenswrapper[4931]: I1129 00:21:31.985457 4931 generic.go:334] "Generic (PLEG): container finished" podID="53ae7271-91fe-4390-a8ff-b39949d8675a" containerID="da47ddcf5db4bd413ff292ed16e42b6fbdcea7fac0fd2eb7ef57aecf5b1f054e" exitCode=0 Nov 29 00:21:31 crc kubenswrapper[4931]: I1129 00:21:31.987887 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"53ae7271-91fe-4390-a8ff-b39949d8675a","Type":"ContainerDied","Data":"da47ddcf5db4bd413ff292ed16e42b6fbdcea7fac0fd2eb7ef57aecf5b1f054e"} Nov 29 00:21:31 crc kubenswrapper[4931]: I1129 00:21:31.987961 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"53ae7271-91fe-4390-a8ff-b39949d8675a","Type":"ContainerDied","Data":"d10e6f41777c1e9d3e95dcd37b6ae19a2ec0a10713f7259aa34609ae5d243135"} Nov 29 00:21:31 crc kubenswrapper[4931]: I1129 00:21:31.987975 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d10e6f41777c1e9d3e95dcd37b6ae19a2ec0a10713f7259aa34609ae5d243135" Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.013298 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.013276233 podStartE2EDuration="2.013276233s" podCreationTimestamp="2025-11-29 00:21:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:21:32.006188788 +0000 UTC m=+1305.168082030" watchObservedRunningTime="2025-11-29 00:21:32.013276233 +0000 UTC m=+1305.175169465" Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.031298 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.165737 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ae7271-91fe-4390-a8ff-b39949d8675a-logs\") pod \"53ae7271-91fe-4390-a8ff-b39949d8675a\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.165855 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ae7271-91fe-4390-a8ff-b39949d8675a-combined-ca-bundle\") pod \"53ae7271-91fe-4390-a8ff-b39949d8675a\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.165964 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53ae7271-91fe-4390-a8ff-b39949d8675a-config-data\") pod \"53ae7271-91fe-4390-a8ff-b39949d8675a\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.166041 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glccm\" (UniqueName: \"kubernetes.io/projected/53ae7271-91fe-4390-a8ff-b39949d8675a-kube-api-access-glccm\") pod \"53ae7271-91fe-4390-a8ff-b39949d8675a\" (UID: \"53ae7271-91fe-4390-a8ff-b39949d8675a\") " Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.167454 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53ae7271-91fe-4390-a8ff-b39949d8675a-logs" (OuterVolumeSpecName: "logs") pod "53ae7271-91fe-4390-a8ff-b39949d8675a" (UID: "53ae7271-91fe-4390-a8ff-b39949d8675a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.172739 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53ae7271-91fe-4390-a8ff-b39949d8675a-kube-api-access-glccm" (OuterVolumeSpecName: "kube-api-access-glccm") pod "53ae7271-91fe-4390-a8ff-b39949d8675a" (UID: "53ae7271-91fe-4390-a8ff-b39949d8675a"). InnerVolumeSpecName "kube-api-access-glccm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.196890 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53ae7271-91fe-4390-a8ff-b39949d8675a-config-data" (OuterVolumeSpecName: "config-data") pod "53ae7271-91fe-4390-a8ff-b39949d8675a" (UID: "53ae7271-91fe-4390-a8ff-b39949d8675a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.200281 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53ae7271-91fe-4390-a8ff-b39949d8675a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53ae7271-91fe-4390-a8ff-b39949d8675a" (UID: "53ae7271-91fe-4390-a8ff-b39949d8675a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.268398 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53ae7271-91fe-4390-a8ff-b39949d8675a-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.268437 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glccm\" (UniqueName: \"kubernetes.io/projected/53ae7271-91fe-4390-a8ff-b39949d8675a-kube-api-access-glccm\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.268453 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ae7271-91fe-4390-a8ff-b39949d8675a-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.268464 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ae7271-91fe-4390-a8ff-b39949d8675a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:32 crc kubenswrapper[4931]: I1129 00:21:32.997781 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.043265 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.055490 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.072494 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 29 00:21:33 crc kubenswrapper[4931]: E1129 00:21:33.074216 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ae7271-91fe-4390-a8ff-b39949d8675a" containerName="nova-api-log" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.074337 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ae7271-91fe-4390-a8ff-b39949d8675a" containerName="nova-api-log" Nov 29 00:21:33 crc kubenswrapper[4931]: E1129 00:21:33.074454 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ae7271-91fe-4390-a8ff-b39949d8675a" containerName="nova-api-api" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.074546 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ae7271-91fe-4390-a8ff-b39949d8675a" containerName="nova-api-api" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.074919 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="53ae7271-91fe-4390-a8ff-b39949d8675a" containerName="nova-api-log" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.076112 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="53ae7271-91fe-4390-a8ff-b39949d8675a" containerName="nova-api-api" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.081509 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.085353 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.096386 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.190088 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133e13c5-0134-4e8a-8d71-5abbe9589368-config-data\") pod \"nova-api-0\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.191038 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133e13c5-0134-4e8a-8d71-5abbe9589368-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.191401 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/133e13c5-0134-4e8a-8d71-5abbe9589368-logs\") pod \"nova-api-0\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.191802 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m4g9\" (UniqueName: \"kubernetes.io/projected/133e13c5-0134-4e8a-8d71-5abbe9589368-kube-api-access-6m4g9\") pod \"nova-api-0\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.227400 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53ae7271-91fe-4390-a8ff-b39949d8675a" path="/var/lib/kubelet/pods/53ae7271-91fe-4390-a8ff-b39949d8675a/volumes" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.293356 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133e13c5-0134-4e8a-8d71-5abbe9589368-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.293440 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/133e13c5-0134-4e8a-8d71-5abbe9589368-logs\") pod \"nova-api-0\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.293544 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m4g9\" (UniqueName: \"kubernetes.io/projected/133e13c5-0134-4e8a-8d71-5abbe9589368-kube-api-access-6m4g9\") pod \"nova-api-0\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.293583 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133e13c5-0134-4e8a-8d71-5abbe9589368-config-data\") pod \"nova-api-0\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.293876 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/133e13c5-0134-4e8a-8d71-5abbe9589368-logs\") pod \"nova-api-0\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.298315 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133e13c5-0134-4e8a-8d71-5abbe9589368-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.305719 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133e13c5-0134-4e8a-8d71-5abbe9589368-config-data\") pod \"nova-api-0\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.321647 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m4g9\" (UniqueName: \"kubernetes.io/projected/133e13c5-0134-4e8a-8d71-5abbe9589368-kube-api-access-6m4g9\") pod \"nova-api-0\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.419481 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:21:33 crc kubenswrapper[4931]: I1129 00:21:33.891081 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:21:34 crc kubenswrapper[4931]: I1129 00:21:34.007544 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"133e13c5-0134-4e8a-8d71-5abbe9589368","Type":"ContainerStarted","Data":"79fdbcd3ffaddf11769a5743f84e98e5a1ef36e0b2500cadd966bed730657fc2"} Nov 29 00:21:34 crc kubenswrapper[4931]: I1129 00:21:34.456367 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 29 00:21:34 crc kubenswrapper[4931]: I1129 00:21:34.526147 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 29 00:21:34 crc kubenswrapper[4931]: I1129 00:21:34.526484 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 29 00:21:35 crc kubenswrapper[4931]: I1129 00:21:35.016697 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"133e13c5-0134-4e8a-8d71-5abbe9589368","Type":"ContainerStarted","Data":"e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed"} Nov 29 00:21:35 crc kubenswrapper[4931]: I1129 00:21:35.016744 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"133e13c5-0134-4e8a-8d71-5abbe9589368","Type":"ContainerStarted","Data":"17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd"} Nov 29 00:21:35 crc kubenswrapper[4931]: I1129 00:21:35.037267 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.037246113 podStartE2EDuration="2.037246113s" podCreationTimestamp="2025-11-29 00:21:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:21:35.035107782 +0000 UTC m=+1308.197001024" watchObservedRunningTime="2025-11-29 00:21:35.037246113 +0000 UTC m=+1308.199139365" Nov 29 00:21:35 crc kubenswrapper[4931]: I1129 00:21:35.468992 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 29 00:21:36 crc kubenswrapper[4931]: I1129 00:21:36.880226 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 29 00:21:39 crc kubenswrapper[4931]: I1129 00:21:39.526694 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 29 00:21:39 crc kubenswrapper[4931]: I1129 00:21:39.527484 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 29 00:21:40 crc kubenswrapper[4931]: I1129 00:21:40.468863 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 29 00:21:40 crc kubenswrapper[4931]: I1129 00:21:40.515255 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 29 00:21:40 crc kubenswrapper[4931]: I1129 00:21:40.539058 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 00:21:40 crc kubenswrapper[4931]: I1129 00:21:40.539117 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 00:21:41 crc kubenswrapper[4931]: I1129 00:21:41.124777 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 29 00:21:43 crc kubenswrapper[4931]: I1129 00:21:43.420486 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 00:21:43 crc kubenswrapper[4931]: I1129 00:21:43.421011 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 00:21:44 crc kubenswrapper[4931]: I1129 00:21:44.503011 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="133e13c5-0134-4e8a-8d71-5abbe9589368" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 29 00:21:44 crc kubenswrapper[4931]: I1129 00:21:44.503011 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="133e13c5-0134-4e8a-8d71-5abbe9589368" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 29 00:21:49 crc kubenswrapper[4931]: I1129 00:21:49.532722 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 29 00:21:49 crc kubenswrapper[4931]: I1129 00:21:49.533651 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 29 00:21:49 crc kubenswrapper[4931]: I1129 00:21:49.537923 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 29 00:21:50 crc kubenswrapper[4931]: I1129 00:21:50.174999 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 29 00:21:50 crc kubenswrapper[4931]: E1129 00:21:50.210917 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-storage-0" podUID="a720d879-8b09-452d-8371-9bf75f28cbde" Nov 29 00:21:51 crc kubenswrapper[4931]: I1129 00:21:51.176069 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.188260 4931 generic.go:334] "Generic (PLEG): container finished" podID="03c88d33-e499-4955-8976-b20c2bd4a09a" containerID="28691ee7f4e971382326216c344002bbde7ae7ff76122bcb79215379b553dedf" exitCode=137 Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.188363 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"03c88d33-e499-4955-8976-b20c2bd4a09a","Type":"ContainerDied","Data":"28691ee7f4e971382326216c344002bbde7ae7ff76122bcb79215379b553dedf"} Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.190206 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"03c88d33-e499-4955-8976-b20c2bd4a09a","Type":"ContainerDied","Data":"a72161fbce2a015f0fbc6613500e29ba6242f665fbd0c360d16711a25606ba80"} Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.190245 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a72161fbce2a015f0fbc6613500e29ba6242f665fbd0c360d16711a25606ba80" Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.263460 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.386637 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03c88d33-e499-4955-8976-b20c2bd4a09a-config-data\") pod \"03c88d33-e499-4955-8976-b20c2bd4a09a\" (UID: \"03c88d33-e499-4955-8976-b20c2bd4a09a\") " Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.386688 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvj4p\" (UniqueName: \"kubernetes.io/projected/03c88d33-e499-4955-8976-b20c2bd4a09a-kube-api-access-qvj4p\") pod \"03c88d33-e499-4955-8976-b20c2bd4a09a\" (UID: \"03c88d33-e499-4955-8976-b20c2bd4a09a\") " Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.386736 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03c88d33-e499-4955-8976-b20c2bd4a09a-combined-ca-bundle\") pod \"03c88d33-e499-4955-8976-b20c2bd4a09a\" (UID: \"03c88d33-e499-4955-8976-b20c2bd4a09a\") " Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.393743 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03c88d33-e499-4955-8976-b20c2bd4a09a-kube-api-access-qvj4p" (OuterVolumeSpecName: "kube-api-access-qvj4p") pod "03c88d33-e499-4955-8976-b20c2bd4a09a" (UID: "03c88d33-e499-4955-8976-b20c2bd4a09a"). InnerVolumeSpecName "kube-api-access-qvj4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.420476 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03c88d33-e499-4955-8976-b20c2bd4a09a-config-data" (OuterVolumeSpecName: "config-data") pod "03c88d33-e499-4955-8976-b20c2bd4a09a" (UID: "03c88d33-e499-4955-8976-b20c2bd4a09a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.425636 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03c88d33-e499-4955-8976-b20c2bd4a09a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03c88d33-e499-4955-8976-b20c2bd4a09a" (UID: "03c88d33-e499-4955-8976-b20c2bd4a09a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.489179 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03c88d33-e499-4955-8976-b20c2bd4a09a-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.489214 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvj4p\" (UniqueName: \"kubernetes.io/projected/03c88d33-e499-4955-8976-b20c2bd4a09a-kube-api-access-qvj4p\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:52 crc kubenswrapper[4931]: I1129 00:21:52.489229 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03c88d33-e499-4955-8976-b20c2bd4a09a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.199828 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.256618 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.277888 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.289691 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 00:21:53 crc kubenswrapper[4931]: E1129 00:21:53.290452 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03c88d33-e499-4955-8976-b20c2bd4a09a" containerName="nova-cell1-novncproxy-novncproxy" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.290470 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="03c88d33-e499-4955-8976-b20c2bd4a09a" containerName="nova-cell1-novncproxy-novncproxy" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.290675 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="03c88d33-e499-4955-8976-b20c2bd4a09a" containerName="nova-cell1-novncproxy-novncproxy" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.291319 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.296419 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.296514 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.296595 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.303779 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.405898 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ad9af8-53d9-48fc-b251-3b580018e51d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.405937 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ad9af8-53d9-48fc-b251-3b580018e51d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.405989 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/82ad9af8-53d9-48fc-b251-3b580018e51d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.406208 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/82ad9af8-53d9-48fc-b251-3b580018e51d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.406259 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzcns\" (UniqueName: \"kubernetes.io/projected/82ad9af8-53d9-48fc-b251-3b580018e51d-kube-api-access-bzcns\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.425306 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.427308 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.429516 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.430735 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.508414 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/82ad9af8-53d9-48fc-b251-3b580018e51d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.509612 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzcns\" (UniqueName: \"kubernetes.io/projected/82ad9af8-53d9-48fc-b251-3b580018e51d-kube-api-access-bzcns\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.509640 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/82ad9af8-53d9-48fc-b251-3b580018e51d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.510920 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ad9af8-53d9-48fc-b251-3b580018e51d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.511301 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ad9af8-53d9-48fc-b251-3b580018e51d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.514052 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/82ad9af8-53d9-48fc-b251-3b580018e51d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.514698 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/82ad9af8-53d9-48fc-b251-3b580018e51d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.517775 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ad9af8-53d9-48fc-b251-3b580018e51d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.524090 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ad9af8-53d9-48fc-b251-3b580018e51d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.525364 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzcns\" (UniqueName: \"kubernetes.io/projected/82ad9af8-53d9-48fc-b251-3b580018e51d-kube-api-access-bzcns\") pod \"nova-cell1-novncproxy-0\" (UID: \"82ad9af8-53d9-48fc-b251-3b580018e51d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:53 crc kubenswrapper[4931]: I1129 00:21:53.615169 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.025118 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:21:54 crc kubenswrapper[4931]: E1129 00:21:54.025386 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:21:54 crc kubenswrapper[4931]: E1129 00:21:54.025459 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 29 00:21:54 crc kubenswrapper[4931]: E1129 00:21:54.025744 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift podName:a720d879-8b09-452d-8371-9bf75f28cbde nodeName:}" failed. No retries permitted until 2025-11-29 00:23:56.025716823 +0000 UTC m=+1449.187610065 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift") pod "swift-storage-0" (UID: "a720d879-8b09-452d-8371-9bf75f28cbde") : configmap "swift-ring-files" not found Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.102574 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 29 00:21:54 crc kubenswrapper[4931]: W1129 00:21:54.106562 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82ad9af8_53d9_48fc_b251_3b580018e51d.slice/crio-e77759c7a0b953b5b474f90e1f84c198fcbf6319e6b140bb54c33cc79b479842 WatchSource:0}: Error finding container e77759c7a0b953b5b474f90e1f84c198fcbf6319e6b140bb54c33cc79b479842: Status 404 returned error can't find the container with id e77759c7a0b953b5b474f90e1f84c198fcbf6319e6b140bb54c33cc79b479842 Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.209926 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82ad9af8-53d9-48fc-b251-3b580018e51d","Type":"ContainerStarted","Data":"e77759c7a0b953b5b474f90e1f84c198fcbf6319e6b140bb54c33cc79b479842"} Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.210276 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.214042 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.359191 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5459cb87c-zlwt7"] Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.361120 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.374901 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5459cb87c-zlwt7"] Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.438285 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-config\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.438353 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x788c\" (UniqueName: \"kubernetes.io/projected/7cc674cd-2e25-45be-acad-b6031e34bd24-kube-api-access-x788c\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.438437 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-ovsdbserver-sb\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.438494 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-dns-svc\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.438568 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-ovsdbserver-nb\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.540334 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-config\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.540402 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x788c\" (UniqueName: \"kubernetes.io/projected/7cc674cd-2e25-45be-acad-b6031e34bd24-kube-api-access-x788c\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.540828 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-ovsdbserver-sb\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.540898 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-dns-svc\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.540985 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-ovsdbserver-nb\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.541383 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-config\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.541626 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-ovsdbserver-sb\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.541821 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-ovsdbserver-nb\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.542074 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-dns-svc\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.564723 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x788c\" (UniqueName: \"kubernetes.io/projected/7cc674cd-2e25-45be-acad-b6031e34bd24-kube-api-access-x788c\") pod \"dnsmasq-dns-5459cb87c-zlwt7\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:54 crc kubenswrapper[4931]: I1129 00:21:54.693996 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:55 crc kubenswrapper[4931]: I1129 00:21:55.204417 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5459cb87c-zlwt7"] Nov 29 00:21:55 crc kubenswrapper[4931]: I1129 00:21:55.236239 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03c88d33-e499-4955-8976-b20c2bd4a09a" path="/var/lib/kubelet/pods/03c88d33-e499-4955-8976-b20c2bd4a09a/volumes" Nov 29 00:21:55 crc kubenswrapper[4931]: I1129 00:21:55.236958 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" event={"ID":"7cc674cd-2e25-45be-acad-b6031e34bd24","Type":"ContainerStarted","Data":"712e1de4433955f789576af7e20c641d1b577ee30ceba6af9aff8d6c385da297"} Nov 29 00:21:55 crc kubenswrapper[4931]: I1129 00:21:55.248549 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82ad9af8-53d9-48fc-b251-3b580018e51d","Type":"ContainerStarted","Data":"2a39e1d747ee71f184d6141feb7d7ead7b9e17831ccbe7528f54706dde7073ce"} Nov 29 00:21:55 crc kubenswrapper[4931]: I1129 00:21:55.270711 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.270674078 podStartE2EDuration="2.270674078s" podCreationTimestamp="2025-11-29 00:21:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:21:55.267610079 +0000 UTC m=+1328.429503321" watchObservedRunningTime="2025-11-29 00:21:55.270674078 +0000 UTC m=+1328.432567310" Nov 29 00:21:56 crc kubenswrapper[4931]: I1129 00:21:56.256149 4931 generic.go:334] "Generic (PLEG): container finished" podID="7cc674cd-2e25-45be-acad-b6031e34bd24" containerID="2138484b45b0f7b2da55011d8546db5b8dc8893b5a2e275c10ca97b22e65745a" exitCode=0 Nov 29 00:21:56 crc kubenswrapper[4931]: I1129 00:21:56.256223 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" event={"ID":"7cc674cd-2e25-45be-acad-b6031e34bd24","Type":"ContainerDied","Data":"2138484b45b0f7b2da55011d8546db5b8dc8893b5a2e275c10ca97b22e65745a"} Nov 29 00:21:56 crc kubenswrapper[4931]: I1129 00:21:56.537826 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:21:56 crc kubenswrapper[4931]: I1129 00:21:56.645870 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:21:56 crc kubenswrapper[4931]: I1129 00:21:56.646222 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="sg-core" containerID="cri-o://c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee" gracePeriod=30 Nov 29 00:21:56 crc kubenswrapper[4931]: I1129 00:21:56.646300 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="ceilometer-notification-agent" containerID="cri-o://22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886" gracePeriod=30 Nov 29 00:21:56 crc kubenswrapper[4931]: I1129 00:21:56.646228 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="proxy-httpd" containerID="cri-o://7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62" gracePeriod=30 Nov 29 00:21:56 crc kubenswrapper[4931]: I1129 00:21:56.646474 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="ceilometer-central-agent" containerID="cri-o://1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557" gracePeriod=30 Nov 29 00:21:57 crc kubenswrapper[4931]: I1129 00:21:57.276391 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" event={"ID":"7cc674cd-2e25-45be-acad-b6031e34bd24","Type":"ContainerStarted","Data":"b13f34346ffb8d9a8b28f94175c7e06620a285a936c46c094f87b08bf6286b72"} Nov 29 00:21:57 crc kubenswrapper[4931]: I1129 00:21:57.278650 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:21:57 crc kubenswrapper[4931]: I1129 00:21:57.282443 4931 generic.go:334] "Generic (PLEG): container finished" podID="a7f18517-445a-4295-90f0-7333785879f2" containerID="7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62" exitCode=0 Nov 29 00:21:57 crc kubenswrapper[4931]: I1129 00:21:57.282479 4931 generic.go:334] "Generic (PLEG): container finished" podID="a7f18517-445a-4295-90f0-7333785879f2" containerID="c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee" exitCode=2 Nov 29 00:21:57 crc kubenswrapper[4931]: I1129 00:21:57.282487 4931 generic.go:334] "Generic (PLEG): container finished" podID="a7f18517-445a-4295-90f0-7333785879f2" containerID="1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557" exitCode=0 Nov 29 00:21:57 crc kubenswrapper[4931]: I1129 00:21:57.282727 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="133e13c5-0134-4e8a-8d71-5abbe9589368" containerName="nova-api-log" containerID="cri-o://17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd" gracePeriod=30 Nov 29 00:21:57 crc kubenswrapper[4931]: I1129 00:21:57.282920 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7f18517-445a-4295-90f0-7333785879f2","Type":"ContainerDied","Data":"7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62"} Nov 29 00:21:57 crc kubenswrapper[4931]: I1129 00:21:57.282975 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7f18517-445a-4295-90f0-7333785879f2","Type":"ContainerDied","Data":"c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee"} Nov 29 00:21:57 crc kubenswrapper[4931]: I1129 00:21:57.282992 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7f18517-445a-4295-90f0-7333785879f2","Type":"ContainerDied","Data":"1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557"} Nov 29 00:21:57 crc kubenswrapper[4931]: I1129 00:21:57.283003 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="133e13c5-0134-4e8a-8d71-5abbe9589368" containerName="nova-api-api" containerID="cri-o://e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed" gracePeriod=30 Nov 29 00:21:57 crc kubenswrapper[4931]: I1129 00:21:57.307618 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" podStartSLOduration=3.307589662 podStartE2EDuration="3.307589662s" podCreationTimestamp="2025-11-29 00:21:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:21:57.297207342 +0000 UTC m=+1330.459100584" watchObservedRunningTime="2025-11-29 00:21:57.307589662 +0000 UTC m=+1330.469482914" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.058886 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.121529 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7f18517-445a-4295-90f0-7333785879f2-run-httpd\") pod \"a7f18517-445a-4295-90f0-7333785879f2\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.121583 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjt6g\" (UniqueName: \"kubernetes.io/projected/a7f18517-445a-4295-90f0-7333785879f2-kube-api-access-bjt6g\") pod \"a7f18517-445a-4295-90f0-7333785879f2\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.121668 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-combined-ca-bundle\") pod \"a7f18517-445a-4295-90f0-7333785879f2\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.121721 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-ceilometer-tls-certs\") pod \"a7f18517-445a-4295-90f0-7333785879f2\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.121746 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7f18517-445a-4295-90f0-7333785879f2-log-httpd\") pod \"a7f18517-445a-4295-90f0-7333785879f2\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.121862 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-config-data\") pod \"a7f18517-445a-4295-90f0-7333785879f2\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.121898 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-sg-core-conf-yaml\") pod \"a7f18517-445a-4295-90f0-7333785879f2\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.121956 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-scripts\") pod \"a7f18517-445a-4295-90f0-7333785879f2\" (UID: \"a7f18517-445a-4295-90f0-7333785879f2\") " Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.122210 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7f18517-445a-4295-90f0-7333785879f2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a7f18517-445a-4295-90f0-7333785879f2" (UID: "a7f18517-445a-4295-90f0-7333785879f2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.122325 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7f18517-445a-4295-90f0-7333785879f2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a7f18517-445a-4295-90f0-7333785879f2" (UID: "a7f18517-445a-4295-90f0-7333785879f2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.122854 4931 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7f18517-445a-4295-90f0-7333785879f2-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.122876 4931 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7f18517-445a-4295-90f0-7333785879f2-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.128535 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-scripts" (OuterVolumeSpecName: "scripts") pod "a7f18517-445a-4295-90f0-7333785879f2" (UID: "a7f18517-445a-4295-90f0-7333785879f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.140373 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f18517-445a-4295-90f0-7333785879f2-kube-api-access-bjt6g" (OuterVolumeSpecName: "kube-api-access-bjt6g") pod "a7f18517-445a-4295-90f0-7333785879f2" (UID: "a7f18517-445a-4295-90f0-7333785879f2"). InnerVolumeSpecName "kube-api-access-bjt6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.169746 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a7f18517-445a-4295-90f0-7333785879f2" (UID: "a7f18517-445a-4295-90f0-7333785879f2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.201265 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a7f18517-445a-4295-90f0-7333785879f2" (UID: "a7f18517-445a-4295-90f0-7333785879f2"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.217169 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7f18517-445a-4295-90f0-7333785879f2" (UID: "a7f18517-445a-4295-90f0-7333785879f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.224323 4931 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.224360 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.224372 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjt6g\" (UniqueName: \"kubernetes.io/projected/a7f18517-445a-4295-90f0-7333785879f2-kube-api-access-bjt6g\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.224387 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.224398 4931 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.255839 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-config-data" (OuterVolumeSpecName: "config-data") pod "a7f18517-445a-4295-90f0-7333785879f2" (UID: "a7f18517-445a-4295-90f0-7333785879f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.299631 4931 generic.go:334] "Generic (PLEG): container finished" podID="a7f18517-445a-4295-90f0-7333785879f2" containerID="22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886" exitCode=0 Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.299706 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7f18517-445a-4295-90f0-7333785879f2","Type":"ContainerDied","Data":"22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886"} Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.299730 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7f18517-445a-4295-90f0-7333785879f2","Type":"ContainerDied","Data":"d22701b6c431eeadbd94fb4cad4f4ce1aec9173e32dcc35f358e9c496d36a4e5"} Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.299760 4931 scope.go:117] "RemoveContainer" containerID="7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.299915 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.304591 4931 generic.go:334] "Generic (PLEG): container finished" podID="133e13c5-0134-4e8a-8d71-5abbe9589368" containerID="17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd" exitCode=143 Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.305214 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"133e13c5-0134-4e8a-8d71-5abbe9589368","Type":"ContainerDied","Data":"17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd"} Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.321908 4931 scope.go:117] "RemoveContainer" containerID="c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.333266 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7f18517-445a-4295-90f0-7333785879f2-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.343890 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.372050 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.374170 4931 scope.go:117] "RemoveContainer" containerID="22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.382229 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:21:58 crc kubenswrapper[4931]: E1129 00:21:58.382626 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="proxy-httpd" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.382645 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="proxy-httpd" Nov 29 00:21:58 crc kubenswrapper[4931]: E1129 00:21:58.382653 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="sg-core" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.382660 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="sg-core" Nov 29 00:21:58 crc kubenswrapper[4931]: E1129 00:21:58.382674 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="ceilometer-central-agent" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.382680 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="ceilometer-central-agent" Nov 29 00:21:58 crc kubenswrapper[4931]: E1129 00:21:58.382693 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="ceilometer-notification-agent" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.382698 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="ceilometer-notification-agent" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.382912 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="sg-core" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.382932 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="proxy-httpd" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.382952 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="ceilometer-notification-agent" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.382960 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f18517-445a-4295-90f0-7333785879f2" containerName="ceilometer-central-agent" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.384634 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.386596 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.386887 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.387042 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.391072 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.420398 4931 scope.go:117] "RemoveContainer" containerID="1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.434655 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaaadf5c-f464-45f6-84d3-5a293743189d-log-httpd\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.434725 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.434758 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-scripts\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.434857 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.434902 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-config-data\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.434920 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ttdd\" (UniqueName: \"kubernetes.io/projected/eaaadf5c-f464-45f6-84d3-5a293743189d-kube-api-access-9ttdd\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.434937 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaaadf5c-f464-45f6-84d3-5a293743189d-run-httpd\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.434951 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.444746 4931 scope.go:117] "RemoveContainer" containerID="7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62" Nov 29 00:21:58 crc kubenswrapper[4931]: E1129 00:21:58.445198 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62\": container with ID starting with 7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62 not found: ID does not exist" containerID="7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.445250 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62"} err="failed to get container status \"7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62\": rpc error: code = NotFound desc = could not find container \"7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62\": container with ID starting with 7d8b4d53726004db8e3b38097304802fa7aa03b199d0f984305ff2531ba1ab62 not found: ID does not exist" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.445277 4931 scope.go:117] "RemoveContainer" containerID="c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee" Nov 29 00:21:58 crc kubenswrapper[4931]: E1129 00:21:58.445710 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee\": container with ID starting with c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee not found: ID does not exist" containerID="c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.445742 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee"} err="failed to get container status \"c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee\": rpc error: code = NotFound desc = could not find container \"c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee\": container with ID starting with c7273db987c6c0f7f6b495b89b25d8a578460782cd201674a04d47bc49a543ee not found: ID does not exist" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.445761 4931 scope.go:117] "RemoveContainer" containerID="22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886" Nov 29 00:21:58 crc kubenswrapper[4931]: E1129 00:21:58.446097 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886\": container with ID starting with 22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886 not found: ID does not exist" containerID="22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.446147 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886"} err="failed to get container status \"22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886\": rpc error: code = NotFound desc = could not find container \"22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886\": container with ID starting with 22993bb0aeafd90c712d1c0e302d14b99a7d2e42b73a88c08947236527ced886 not found: ID does not exist" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.446179 4931 scope.go:117] "RemoveContainer" containerID="1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557" Nov 29 00:21:58 crc kubenswrapper[4931]: E1129 00:21:58.446536 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557\": container with ID starting with 1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557 not found: ID does not exist" containerID="1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.446567 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557"} err="failed to get container status \"1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557\": rpc error: code = NotFound desc = could not find container \"1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557\": container with ID starting with 1c5d22d3b4b666c7f6fbf77b9d88779ac3f9b94bf9f64461900cd7569fc19557 not found: ID does not exist" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.515982 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:21:58 crc kubenswrapper[4931]: E1129 00:21:58.516739 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-9ttdd log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="eaaadf5c-f464-45f6-84d3-5a293743189d" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.536245 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaaadf5c-f464-45f6-84d3-5a293743189d-log-httpd\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.536325 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.536362 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-scripts\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.536408 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.536458 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-config-data\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.536484 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ttdd\" (UniqueName: \"kubernetes.io/projected/eaaadf5c-f464-45f6-84d3-5a293743189d-kube-api-access-9ttdd\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.536505 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaaadf5c-f464-45f6-84d3-5a293743189d-run-httpd\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.536994 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.537303 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaaadf5c-f464-45f6-84d3-5a293743189d-run-httpd\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.537836 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaaadf5c-f464-45f6-84d3-5a293743189d-log-httpd\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.542151 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-scripts\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.542390 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-config-data\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.542878 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.544134 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.544203 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.553532 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ttdd\" (UniqueName: \"kubernetes.io/projected/eaaadf5c-f464-45f6-84d3-5a293743189d-kube-api-access-9ttdd\") pod \"ceilometer-0\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " pod="openstack/ceilometer-0" Nov 29 00:21:58 crc kubenswrapper[4931]: I1129 00:21:58.615403 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.226832 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7f18517-445a-4295-90f0-7333785879f2" path="/var/lib/kubelet/pods/a7f18517-445a-4295-90f0-7333785879f2/volumes" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.315671 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.325645 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.456000 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ttdd\" (UniqueName: \"kubernetes.io/projected/eaaadf5c-f464-45f6-84d3-5a293743189d-kube-api-access-9ttdd\") pod \"eaaadf5c-f464-45f6-84d3-5a293743189d\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.456046 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-ceilometer-tls-certs\") pod \"eaaadf5c-f464-45f6-84d3-5a293743189d\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.456085 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-config-data\") pod \"eaaadf5c-f464-45f6-84d3-5a293743189d\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.456176 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-scripts\") pod \"eaaadf5c-f464-45f6-84d3-5a293743189d\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.456204 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-sg-core-conf-yaml\") pod \"eaaadf5c-f464-45f6-84d3-5a293743189d\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.456261 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaaadf5c-f464-45f6-84d3-5a293743189d-run-httpd\") pod \"eaaadf5c-f464-45f6-84d3-5a293743189d\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.456301 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaaadf5c-f464-45f6-84d3-5a293743189d-log-httpd\") pod \"eaaadf5c-f464-45f6-84d3-5a293743189d\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.456327 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-combined-ca-bundle\") pod \"eaaadf5c-f464-45f6-84d3-5a293743189d\" (UID: \"eaaadf5c-f464-45f6-84d3-5a293743189d\") " Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.481048 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaaadf5c-f464-45f6-84d3-5a293743189d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eaaadf5c-f464-45f6-84d3-5a293743189d" (UID: "eaaadf5c-f464-45f6-84d3-5a293743189d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.486306 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaaadf5c-f464-45f6-84d3-5a293743189d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eaaadf5c-f464-45f6-84d3-5a293743189d" (UID: "eaaadf5c-f464-45f6-84d3-5a293743189d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.487851 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-config-data" (OuterVolumeSpecName: "config-data") pod "eaaadf5c-f464-45f6-84d3-5a293743189d" (UID: "eaaadf5c-f464-45f6-84d3-5a293743189d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.490389 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "eaaadf5c-f464-45f6-84d3-5a293743189d" (UID: "eaaadf5c-f464-45f6-84d3-5a293743189d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.490499 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eaaadf5c-f464-45f6-84d3-5a293743189d" (UID: "eaaadf5c-f464-45f6-84d3-5a293743189d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.507003 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-scripts" (OuterVolumeSpecName: "scripts") pod "eaaadf5c-f464-45f6-84d3-5a293743189d" (UID: "eaaadf5c-f464-45f6-84d3-5a293743189d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.507055 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaaadf5c-f464-45f6-84d3-5a293743189d-kube-api-access-9ttdd" (OuterVolumeSpecName: "kube-api-access-9ttdd") pod "eaaadf5c-f464-45f6-84d3-5a293743189d" (UID: "eaaadf5c-f464-45f6-84d3-5a293743189d"). InnerVolumeSpecName "kube-api-access-9ttdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.507057 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaaadf5c-f464-45f6-84d3-5a293743189d" (UID: "eaaadf5c-f464-45f6-84d3-5a293743189d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.566270 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.566311 4931 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.566324 4931 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaaadf5c-f464-45f6-84d3-5a293743189d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.566336 4931 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaaadf5c-f464-45f6-84d3-5a293743189d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.566349 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.566361 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ttdd\" (UniqueName: \"kubernetes.io/projected/eaaadf5c-f464-45f6-84d3-5a293743189d-kube-api-access-9ttdd\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.566373 4931 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:21:59 crc kubenswrapper[4931]: I1129 00:21:59.566385 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaaadf5c-f464-45f6-84d3-5a293743189d-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.322626 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.399651 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.419982 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.437448 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.440244 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.442002 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.442183 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.442289 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.446928 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.584439 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-scripts\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.584677 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ngrx\" (UniqueName: \"kubernetes.io/projected/3ecafd71-ca29-4382-9d56-d1e7422d113b-kube-api-access-2ngrx\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.584766 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.584827 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.584866 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ecafd71-ca29-4382-9d56-d1e7422d113b-run-httpd\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.584893 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ecafd71-ca29-4382-9d56-d1e7422d113b-log-httpd\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.584919 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.584934 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-config-data\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.689693 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ecafd71-ca29-4382-9d56-d1e7422d113b-run-httpd\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.689752 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ecafd71-ca29-4382-9d56-d1e7422d113b-log-httpd\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.689783 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.689801 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-config-data\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.689862 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-scripts\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.689882 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ngrx\" (UniqueName: \"kubernetes.io/projected/3ecafd71-ca29-4382-9d56-d1e7422d113b-kube-api-access-2ngrx\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.689955 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.690001 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.690170 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ecafd71-ca29-4382-9d56-d1e7422d113b-run-httpd\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.690649 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ecafd71-ca29-4382-9d56-d1e7422d113b-log-httpd\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.695573 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.696002 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.699569 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.699783 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-config-data\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.702052 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ecafd71-ca29-4382-9d56-d1e7422d113b-scripts\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.712730 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ngrx\" (UniqueName: \"kubernetes.io/projected/3ecafd71-ca29-4382-9d56-d1e7422d113b-kube-api-access-2ngrx\") pod \"ceilometer-0\" (UID: \"3ecafd71-ca29-4382-9d56-d1e7422d113b\") " pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.858905 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 29 00:22:00 crc kubenswrapper[4931]: I1129 00:22:00.913041 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.001524 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133e13c5-0134-4e8a-8d71-5abbe9589368-combined-ca-bundle\") pod \"133e13c5-0134-4e8a-8d71-5abbe9589368\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.001875 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m4g9\" (UniqueName: \"kubernetes.io/projected/133e13c5-0134-4e8a-8d71-5abbe9589368-kube-api-access-6m4g9\") pod \"133e13c5-0134-4e8a-8d71-5abbe9589368\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.001938 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133e13c5-0134-4e8a-8d71-5abbe9589368-config-data\") pod \"133e13c5-0134-4e8a-8d71-5abbe9589368\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.001959 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/133e13c5-0134-4e8a-8d71-5abbe9589368-logs\") pod \"133e13c5-0134-4e8a-8d71-5abbe9589368\" (UID: \"133e13c5-0134-4e8a-8d71-5abbe9589368\") " Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.002904 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/133e13c5-0134-4e8a-8d71-5abbe9589368-logs" (OuterVolumeSpecName: "logs") pod "133e13c5-0134-4e8a-8d71-5abbe9589368" (UID: "133e13c5-0134-4e8a-8d71-5abbe9589368"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.011931 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/133e13c5-0134-4e8a-8d71-5abbe9589368-kube-api-access-6m4g9" (OuterVolumeSpecName: "kube-api-access-6m4g9") pod "133e13c5-0134-4e8a-8d71-5abbe9589368" (UID: "133e13c5-0134-4e8a-8d71-5abbe9589368"). InnerVolumeSpecName "kube-api-access-6m4g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.037436 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133e13c5-0134-4e8a-8d71-5abbe9589368-config-data" (OuterVolumeSpecName: "config-data") pod "133e13c5-0134-4e8a-8d71-5abbe9589368" (UID: "133e13c5-0134-4e8a-8d71-5abbe9589368"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.045153 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133e13c5-0134-4e8a-8d71-5abbe9589368-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "133e13c5-0134-4e8a-8d71-5abbe9589368" (UID: "133e13c5-0134-4e8a-8d71-5abbe9589368"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.104178 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133e13c5-0134-4e8a-8d71-5abbe9589368-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.104208 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m4g9\" (UniqueName: \"kubernetes.io/projected/133e13c5-0134-4e8a-8d71-5abbe9589368-kube-api-access-6m4g9\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.104222 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133e13c5-0134-4e8a-8d71-5abbe9589368-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.104232 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/133e13c5-0134-4e8a-8d71-5abbe9589368-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.222450 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaaadf5c-f464-45f6-84d3-5a293743189d" path="/var/lib/kubelet/pods/eaaadf5c-f464-45f6-84d3-5a293743189d/volumes" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.333629 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 29 00:22:01 crc kubenswrapper[4931]: W1129 00:22:01.333708 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ecafd71_ca29_4382_9d56_d1e7422d113b.slice/crio-defa02be5c822af0b2d85e341687861295e90b3f1b17e106aef1e80b352f26b0 WatchSource:0}: Error finding container defa02be5c822af0b2d85e341687861295e90b3f1b17e106aef1e80b352f26b0: Status 404 returned error can't find the container with id defa02be5c822af0b2d85e341687861295e90b3f1b17e106aef1e80b352f26b0 Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.337048 4931 generic.go:334] "Generic (PLEG): container finished" podID="133e13c5-0134-4e8a-8d71-5abbe9589368" containerID="e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed" exitCode=0 Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.337086 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"133e13c5-0134-4e8a-8d71-5abbe9589368","Type":"ContainerDied","Data":"e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed"} Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.337089 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.337109 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"133e13c5-0134-4e8a-8d71-5abbe9589368","Type":"ContainerDied","Data":"79fdbcd3ffaddf11769a5743f84e98e5a1ef36e0b2500cadd966bed730657fc2"} Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.337142 4931 scope.go:117] "RemoveContainer" containerID="e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.337259 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.363157 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.371800 4931 scope.go:117] "RemoveContainer" containerID="17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.379298 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.395505 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 29 00:22:01 crc kubenswrapper[4931]: E1129 00:22:01.396049 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133e13c5-0134-4e8a-8d71-5abbe9589368" containerName="nova-api-api" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.396076 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="133e13c5-0134-4e8a-8d71-5abbe9589368" containerName="nova-api-api" Nov 29 00:22:01 crc kubenswrapper[4931]: E1129 00:22:01.396089 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133e13c5-0134-4e8a-8d71-5abbe9589368" containerName="nova-api-log" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.396136 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="133e13c5-0134-4e8a-8d71-5abbe9589368" containerName="nova-api-log" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.396321 4931 scope.go:117] "RemoveContainer" containerID="e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.396389 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="133e13c5-0134-4e8a-8d71-5abbe9589368" containerName="nova-api-api" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.396413 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="133e13c5-0134-4e8a-8d71-5abbe9589368" containerName="nova-api-log" Nov 29 00:22:01 crc kubenswrapper[4931]: E1129 00:22:01.396701 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed\": container with ID starting with e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed not found: ID does not exist" containerID="e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.396739 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed"} err="failed to get container status \"e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed\": rpc error: code = NotFound desc = could not find container \"e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed\": container with ID starting with e9df8bf7cf6357abb5bf44aee72fe6ef8b8c0e233d43c0da47512d106ba141ed not found: ID does not exist" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.396768 4931 scope.go:117] "RemoveContainer" containerID="17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd" Nov 29 00:22:01 crc kubenswrapper[4931]: E1129 00:22:01.397445 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd\": container with ID starting with 17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd not found: ID does not exist" containerID="17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.397477 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd"} err="failed to get container status \"17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd\": rpc error: code = NotFound desc = could not find container \"17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd\": container with ID starting with 17cb27d34f8786af70681af1590e9da1fdd222721fe28c5393659adf15a50efd not found: ID does not exist" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.398649 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.412194 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.412431 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.412599 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.428287 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.511317 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a0b9809-f5c3-4426-92a6-7cff2a78b936-logs\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.511382 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-public-tls-certs\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.511410 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.511865 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cj25\" (UniqueName: \"kubernetes.io/projected/8a0b9809-f5c3-4426-92a6-7cff2a78b936-kube-api-access-5cj25\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.511943 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-config-data\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.511994 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.613989 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cj25\" (UniqueName: \"kubernetes.io/projected/8a0b9809-f5c3-4426-92a6-7cff2a78b936-kube-api-access-5cj25\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.614371 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-config-data\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.614401 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.615460 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a0b9809-f5c3-4426-92a6-7cff2a78b936-logs\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.615529 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-public-tls-certs\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.615559 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.615892 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a0b9809-f5c3-4426-92a6-7cff2a78b936-logs\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.619352 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-public-tls-certs\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.619413 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.619530 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.620183 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-config-data\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.640216 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cj25\" (UniqueName: \"kubernetes.io/projected/8a0b9809-f5c3-4426-92a6-7cff2a78b936-kube-api-access-5cj25\") pod \"nova-api-0\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " pod="openstack/nova-api-0" Nov 29 00:22:01 crc kubenswrapper[4931]: I1129 00:22:01.736121 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:22:02 crc kubenswrapper[4931]: I1129 00:22:02.202505 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:22:02 crc kubenswrapper[4931]: W1129 00:22:02.207335 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a0b9809_f5c3_4426_92a6_7cff2a78b936.slice/crio-699342d05d5a38b3469f6402e8ab309286b541f1ea56471850248dea17ad93d4 WatchSource:0}: Error finding container 699342d05d5a38b3469f6402e8ab309286b541f1ea56471850248dea17ad93d4: Status 404 returned error can't find the container with id 699342d05d5a38b3469f6402e8ab309286b541f1ea56471850248dea17ad93d4 Nov 29 00:22:02 crc kubenswrapper[4931]: I1129 00:22:02.349399 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ecafd71-ca29-4382-9d56-d1e7422d113b","Type":"ContainerStarted","Data":"347fdc5d17dd281cc4e12e420a58873b1eb47202b0ed83f06b5f39eba7333a0c"} Nov 29 00:22:02 crc kubenswrapper[4931]: I1129 00:22:02.349445 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ecafd71-ca29-4382-9d56-d1e7422d113b","Type":"ContainerStarted","Data":"defa02be5c822af0b2d85e341687861295e90b3f1b17e106aef1e80b352f26b0"} Nov 29 00:22:02 crc kubenswrapper[4931]: I1129 00:22:02.352066 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8a0b9809-f5c3-4426-92a6-7cff2a78b936","Type":"ContainerStarted","Data":"699342d05d5a38b3469f6402e8ab309286b541f1ea56471850248dea17ad93d4"} Nov 29 00:22:03 crc kubenswrapper[4931]: I1129 00:22:03.221990 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="133e13c5-0134-4e8a-8d71-5abbe9589368" path="/var/lib/kubelet/pods/133e13c5-0134-4e8a-8d71-5abbe9589368/volumes" Nov 29 00:22:03 crc kubenswrapper[4931]: I1129 00:22:03.371078 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ecafd71-ca29-4382-9d56-d1e7422d113b","Type":"ContainerStarted","Data":"e504b246273674273dcdd2e0f660efadd197a654edbb3c0df5b4c7ef03e57f14"} Nov 29 00:22:03 crc kubenswrapper[4931]: I1129 00:22:03.373281 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8a0b9809-f5c3-4426-92a6-7cff2a78b936","Type":"ContainerStarted","Data":"a476459aedbdc5c2cf64242efd1af0a97196ac1c973814ab7881d62669e67140"} Nov 29 00:22:03 crc kubenswrapper[4931]: I1129 00:22:03.373310 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8a0b9809-f5c3-4426-92a6-7cff2a78b936","Type":"ContainerStarted","Data":"3d283e14a08e74b7385bdb154bac500e9b7449a59333d8c5669a009b9535ff67"} Nov 29 00:22:03 crc kubenswrapper[4931]: I1129 00:22:03.402647 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.402629548 podStartE2EDuration="2.402629548s" podCreationTimestamp="2025-11-29 00:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:22:03.401719031 +0000 UTC m=+1336.563612273" watchObservedRunningTime="2025-11-29 00:22:03.402629548 +0000 UTC m=+1336.564522780" Nov 29 00:22:03 crc kubenswrapper[4931]: I1129 00:22:03.616136 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:22:03 crc kubenswrapper[4931]: I1129 00:22:03.633526 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.391463 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ecafd71-ca29-4382-9d56-d1e7422d113b","Type":"ContainerStarted","Data":"487c3f0acae553ae27ad09a55d39cf6d785fc4dabd233680e88c6fbb083de382"} Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.412798 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.551090 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-bwdbx"] Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.552789 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.556147 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.556395 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.557873 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bwdbx"] Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.609872 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjfmp\" (UniqueName: \"kubernetes.io/projected/9f619a0b-81e1-4881-abaf-643f69860a61-kube-api-access-gjfmp\") pod \"nova-cell1-cell-mapping-bwdbx\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.609934 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-scripts\") pod \"nova-cell1-cell-mapping-bwdbx\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.610253 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-config-data\") pod \"nova-cell1-cell-mapping-bwdbx\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.610372 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bwdbx\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.696330 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.717702 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-config-data\") pod \"nova-cell1-cell-mapping-bwdbx\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.717789 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bwdbx\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.717853 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjfmp\" (UniqueName: \"kubernetes.io/projected/9f619a0b-81e1-4881-abaf-643f69860a61-kube-api-access-gjfmp\") pod \"nova-cell1-cell-mapping-bwdbx\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.717872 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-scripts\") pod \"nova-cell1-cell-mapping-bwdbx\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.725377 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-scripts\") pod \"nova-cell1-cell-mapping-bwdbx\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.725500 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bwdbx\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.728426 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-config-data\") pod \"nova-cell1-cell-mapping-bwdbx\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.756426 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6ccb6797-cgp75"] Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.756750 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" podUID="149e0e4c-6506-46d0-a6a4-965d8966c71e" containerName="dnsmasq-dns" containerID="cri-o://a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8" gracePeriod=10 Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.768857 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjfmp\" (UniqueName: \"kubernetes.io/projected/9f619a0b-81e1-4881-abaf-643f69860a61-kube-api-access-gjfmp\") pod \"nova-cell1-cell-mapping-bwdbx\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:04 crc kubenswrapper[4931]: I1129 00:22:04.880652 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.295143 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.332327 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-config\") pod \"149e0e4c-6506-46d0-a6a4-965d8966c71e\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.332417 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4kxn\" (UniqueName: \"kubernetes.io/projected/149e0e4c-6506-46d0-a6a4-965d8966c71e-kube-api-access-l4kxn\") pod \"149e0e4c-6506-46d0-a6a4-965d8966c71e\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.332461 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-dns-svc\") pod \"149e0e4c-6506-46d0-a6a4-965d8966c71e\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.332499 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-ovsdbserver-nb\") pod \"149e0e4c-6506-46d0-a6a4-965d8966c71e\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.332528 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-ovsdbserver-sb\") pod \"149e0e4c-6506-46d0-a6a4-965d8966c71e\" (UID: \"149e0e4c-6506-46d0-a6a4-965d8966c71e\") " Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.343082 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/149e0e4c-6506-46d0-a6a4-965d8966c71e-kube-api-access-l4kxn" (OuterVolumeSpecName: "kube-api-access-l4kxn") pod "149e0e4c-6506-46d0-a6a4-965d8966c71e" (UID: "149e0e4c-6506-46d0-a6a4-965d8966c71e"). InnerVolumeSpecName "kube-api-access-l4kxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.407562 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-config" (OuterVolumeSpecName: "config") pod "149e0e4c-6506-46d0-a6a4-965d8966c71e" (UID: "149e0e4c-6506-46d0-a6a4-965d8966c71e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.411729 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "149e0e4c-6506-46d0-a6a4-965d8966c71e" (UID: "149e0e4c-6506-46d0-a6a4-965d8966c71e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.420645 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "149e0e4c-6506-46d0-a6a4-965d8966c71e" (UID: "149e0e4c-6506-46d0-a6a4-965d8966c71e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.429186 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ecafd71-ca29-4382-9d56-d1e7422d113b","Type":"ContainerStarted","Data":"102e494877e380ac86c2e516cb5f3a6685718be9bfabf826c3f3c7f9c739b0d3"} Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.429338 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.433063 4931 generic.go:334] "Generic (PLEG): container finished" podID="149e0e4c-6506-46d0-a6a4-965d8966c71e" containerID="a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8" exitCode=0 Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.433120 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.433112 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "149e0e4c-6506-46d0-a6a4-965d8966c71e" (UID: "149e0e4c-6506-46d0-a6a4-965d8966c71e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.433161 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" event={"ID":"149e0e4c-6506-46d0-a6a4-965d8966c71e","Type":"ContainerDied","Data":"a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8"} Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.433210 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6ccb6797-cgp75" event={"ID":"149e0e4c-6506-46d0-a6a4-965d8966c71e","Type":"ContainerDied","Data":"c599ca4df24ed604688409fdb767083f346f6c4063d4fe3fab6448f726a8bcad"} Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.433227 4931 scope.go:117] "RemoveContainer" containerID="a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.434435 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4kxn\" (UniqueName: \"kubernetes.io/projected/149e0e4c-6506-46d0-a6a4-965d8966c71e-kube-api-access-l4kxn\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.434708 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.434729 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.434742 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.434753 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/149e0e4c-6506-46d0-a6a4-965d8966c71e-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.457288 4931 scope.go:117] "RemoveContainer" containerID="805fc2096b60abcd283157465e921f32416a8cbbb573923e33db514a02922623" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.473049 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.997776863 podStartE2EDuration="5.47303169s" podCreationTimestamp="2025-11-29 00:22:00 +0000 UTC" firstStartedPulling="2025-11-29 00:22:01.336679233 +0000 UTC m=+1334.498572465" lastFinishedPulling="2025-11-29 00:22:04.81193406 +0000 UTC m=+1337.973827292" observedRunningTime="2025-11-29 00:22:05.466033398 +0000 UTC m=+1338.627926650" watchObservedRunningTime="2025-11-29 00:22:05.47303169 +0000 UTC m=+1338.634924922" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.492435 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6ccb6797-cgp75"] Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.501856 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c6ccb6797-cgp75"] Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.502061 4931 scope.go:117] "RemoveContainer" containerID="a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8" Nov 29 00:22:05 crc kubenswrapper[4931]: E1129 00:22:05.502623 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8\": container with ID starting with a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8 not found: ID does not exist" containerID="a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.502670 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8"} err="failed to get container status \"a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8\": rpc error: code = NotFound desc = could not find container \"a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8\": container with ID starting with a6a608dd1160607bc959f5b65e0e0fcf32cf6eb3ecea646dc22cce2b577cadf8 not found: ID does not exist" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.502701 4931 scope.go:117] "RemoveContainer" containerID="805fc2096b60abcd283157465e921f32416a8cbbb573923e33db514a02922623" Nov 29 00:22:05 crc kubenswrapper[4931]: E1129 00:22:05.503221 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"805fc2096b60abcd283157465e921f32416a8cbbb573923e33db514a02922623\": container with ID starting with 805fc2096b60abcd283157465e921f32416a8cbbb573923e33db514a02922623 not found: ID does not exist" containerID="805fc2096b60abcd283157465e921f32416a8cbbb573923e33db514a02922623" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.503251 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"805fc2096b60abcd283157465e921f32416a8cbbb573923e33db514a02922623"} err="failed to get container status \"805fc2096b60abcd283157465e921f32416a8cbbb573923e33db514a02922623\": rpc error: code = NotFound desc = could not find container \"805fc2096b60abcd283157465e921f32416a8cbbb573923e33db514a02922623\": container with ID starting with 805fc2096b60abcd283157465e921f32416a8cbbb573923e33db514a02922623 not found: ID does not exist" Nov 29 00:22:05 crc kubenswrapper[4931]: I1129 00:22:05.528071 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bwdbx"] Nov 29 00:22:06 crc kubenswrapper[4931]: I1129 00:22:06.453694 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bwdbx" event={"ID":"9f619a0b-81e1-4881-abaf-643f69860a61","Type":"ContainerStarted","Data":"ca7d2ede55474b3056f1012b12fe3e926f0e93770a1184197b2197d9a96cc0fc"} Nov 29 00:22:06 crc kubenswrapper[4931]: I1129 00:22:06.454128 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bwdbx" event={"ID":"9f619a0b-81e1-4881-abaf-643f69860a61","Type":"ContainerStarted","Data":"2f252978f1347fe8d77a8badbbc4fe94213454baceda37ceb0468506595af713"} Nov 29 00:22:06 crc kubenswrapper[4931]: I1129 00:22:06.466652 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-bwdbx" podStartSLOduration=2.466639328 podStartE2EDuration="2.466639328s" podCreationTimestamp="2025-11-29 00:22:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:22:06.465314349 +0000 UTC m=+1339.627207581" watchObservedRunningTime="2025-11-29 00:22:06.466639328 +0000 UTC m=+1339.628532550" Nov 29 00:22:07 crc kubenswrapper[4931]: E1129 00:22:07.050924 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-proxy-764cb44577-7q67f" podUID="550e16f7-f101-4a91-8e2f-45af3430d6b9" Nov 29 00:22:07 crc kubenswrapper[4931]: I1129 00:22:07.228769 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="149e0e4c-6506-46d0-a6a4-965d8966c71e" path="/var/lib/kubelet/pods/149e0e4c-6506-46d0-a6a4-965d8966c71e/volumes" Nov 29 00:22:07 crc kubenswrapper[4931]: I1129 00:22:07.465178 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:22:11 crc kubenswrapper[4931]: I1129 00:22:11.497397 4931 generic.go:334] "Generic (PLEG): container finished" podID="9f619a0b-81e1-4881-abaf-643f69860a61" containerID="ca7d2ede55474b3056f1012b12fe3e926f0e93770a1184197b2197d9a96cc0fc" exitCode=0 Nov 29 00:22:11 crc kubenswrapper[4931]: I1129 00:22:11.497473 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bwdbx" event={"ID":"9f619a0b-81e1-4881-abaf-643f69860a61","Type":"ContainerDied","Data":"ca7d2ede55474b3056f1012b12fe3e926f0e93770a1184197b2197d9a96cc0fc"} Nov 29 00:22:11 crc kubenswrapper[4931]: I1129 00:22:11.736877 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 00:22:11 crc kubenswrapper[4931]: I1129 00:22:11.736948 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 00:22:12 crc kubenswrapper[4931]: I1129 00:22:12.186968 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:22:12 crc kubenswrapper[4931]: E1129 00:22:12.187515 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:22:12 crc kubenswrapper[4931]: E1129 00:22:12.187534 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-764cb44577-7q67f: configmap "swift-ring-files" not found Nov 29 00:22:12 crc kubenswrapper[4931]: E1129 00:22:12.187585 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift podName:550e16f7-f101-4a91-8e2f-45af3430d6b9 nodeName:}" failed. No retries permitted until 2025-11-29 00:24:14.187567432 +0000 UTC m=+1467.349460664 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift") pod "swift-proxy-764cb44577-7q67f" (UID: "550e16f7-f101-4a91-8e2f-45af3430d6b9") : configmap "swift-ring-files" not found Nov 29 00:22:12 crc kubenswrapper[4931]: I1129 00:22:12.755002 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 00:22:12 crc kubenswrapper[4931]: I1129 00:22:12.755016 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 00:22:12 crc kubenswrapper[4931]: I1129 00:22:12.849088 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:12 crc kubenswrapper[4931]: I1129 00:22:12.914235 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjfmp\" (UniqueName: \"kubernetes.io/projected/9f619a0b-81e1-4881-abaf-643f69860a61-kube-api-access-gjfmp\") pod \"9f619a0b-81e1-4881-abaf-643f69860a61\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " Nov 29 00:22:12 crc kubenswrapper[4931]: I1129 00:22:12.914303 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-combined-ca-bundle\") pod \"9f619a0b-81e1-4881-abaf-643f69860a61\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " Nov 29 00:22:12 crc kubenswrapper[4931]: I1129 00:22:12.914348 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-config-data\") pod \"9f619a0b-81e1-4881-abaf-643f69860a61\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " Nov 29 00:22:12 crc kubenswrapper[4931]: I1129 00:22:12.914437 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-scripts\") pod \"9f619a0b-81e1-4881-abaf-643f69860a61\" (UID: \"9f619a0b-81e1-4881-abaf-643f69860a61\") " Nov 29 00:22:12 crc kubenswrapper[4931]: I1129 00:22:12.929159 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f619a0b-81e1-4881-abaf-643f69860a61-kube-api-access-gjfmp" (OuterVolumeSpecName: "kube-api-access-gjfmp") pod "9f619a0b-81e1-4881-abaf-643f69860a61" (UID: "9f619a0b-81e1-4881-abaf-643f69860a61"). InnerVolumeSpecName "kube-api-access-gjfmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:22:12 crc kubenswrapper[4931]: I1129 00:22:12.958340 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-scripts" (OuterVolumeSpecName: "scripts") pod "9f619a0b-81e1-4881-abaf-643f69860a61" (UID: "9f619a0b-81e1-4881-abaf-643f69860a61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:12 crc kubenswrapper[4931]: I1129 00:22:12.959964 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f619a0b-81e1-4881-abaf-643f69860a61" (UID: "9f619a0b-81e1-4881-abaf-643f69860a61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:12 crc kubenswrapper[4931]: I1129 00:22:12.962420 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-config-data" (OuterVolumeSpecName: "config-data") pod "9f619a0b-81e1-4881-abaf-643f69860a61" (UID: "9f619a0b-81e1-4881-abaf-643f69860a61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.016036 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.016080 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjfmp\" (UniqueName: \"kubernetes.io/projected/9f619a0b-81e1-4881-abaf-643f69860a61-kube-api-access-gjfmp\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.016096 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.016107 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f619a0b-81e1-4881-abaf-643f69860a61-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.514528 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bwdbx" event={"ID":"9f619a0b-81e1-4881-abaf-643f69860a61","Type":"ContainerDied","Data":"2f252978f1347fe8d77a8badbbc4fe94213454baceda37ceb0468506595af713"} Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.514864 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f252978f1347fe8d77a8badbbc4fe94213454baceda37ceb0468506595af713" Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.514593 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bwdbx" Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.807150 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.808370 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d62bbf08-1341-44a7-b3d2-d85b0490f6f3" containerName="nova-scheduler-scheduler" containerID="cri-o://01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac" gracePeriod=30 Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.819793 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.820067 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" containerName="nova-api-log" containerID="cri-o://3d283e14a08e74b7385bdb154bac500e9b7449a59333d8c5669a009b9535ff67" gracePeriod=30 Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.820538 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" containerName="nova-api-api" containerID="cri-o://a476459aedbdc5c2cf64242efd1af0a97196ac1c973814ab7881d62669e67140" gracePeriod=30 Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.853090 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.853826 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerName="nova-metadata-metadata" containerID="cri-o://88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b" gracePeriod=30 Nov 29 00:22:13 crc kubenswrapper[4931]: I1129 00:22:13.853770 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerName="nova-metadata-log" containerID="cri-o://65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff" gracePeriod=30 Nov 29 00:22:14 crc kubenswrapper[4931]: I1129 00:22:14.525397 4931 generic.go:334] "Generic (PLEG): container finished" podID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" containerID="3d283e14a08e74b7385bdb154bac500e9b7449a59333d8c5669a009b9535ff67" exitCode=143 Nov 29 00:22:14 crc kubenswrapper[4931]: I1129 00:22:14.525443 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8a0b9809-f5c3-4426-92a6-7cff2a78b936","Type":"ContainerDied","Data":"3d283e14a08e74b7385bdb154bac500e9b7449a59333d8c5669a009b9535ff67"} Nov 29 00:22:14 crc kubenswrapper[4931]: I1129 00:22:14.527644 4931 generic.go:334] "Generic (PLEG): container finished" podID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerID="65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff" exitCode=143 Nov 29 00:22:14 crc kubenswrapper[4931]: I1129 00:22:14.527675 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2a7ac81-47a3-4c55-90bc-f19842c87e37","Type":"ContainerDied","Data":"65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff"} Nov 29 00:22:15 crc kubenswrapper[4931]: E1129 00:22:15.469954 4931 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 29 00:22:15 crc kubenswrapper[4931]: E1129 00:22:15.472648 4931 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 29 00:22:15 crc kubenswrapper[4931]: E1129 00:22:15.473976 4931 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 29 00:22:15 crc kubenswrapper[4931]: E1129 00:22:15.474113 4931 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="d62bbf08-1341-44a7-b3d2-d85b0490f6f3" containerName="nova-scheduler-scheduler" Nov 29 00:22:16 crc kubenswrapper[4931]: I1129 00:22:16.993767 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:57870->10.217.0.192:8775: read: connection reset by peer" Nov 29 00:22:16 crc kubenswrapper[4931]: I1129 00:22:16.993767 4931 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:57880->10.217.0.192:8775: read: connection reset by peer" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.504454 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.564316 4931 generic.go:334] "Generic (PLEG): container finished" podID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerID="88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b" exitCode=0 Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.564355 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2a7ac81-47a3-4c55-90bc-f19842c87e37","Type":"ContainerDied","Data":"88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b"} Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.564382 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c2a7ac81-47a3-4c55-90bc-f19842c87e37","Type":"ContainerDied","Data":"8fe866567202b637e05870e2338406b163d6f4f6c65a24ad70f257759765f132"} Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.564398 4931 scope.go:117] "RemoveContainer" containerID="88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.564514 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.586498 4931 scope.go:117] "RemoveContainer" containerID="65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.608152 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-nova-metadata-tls-certs\") pod \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.608205 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9t68\" (UniqueName: \"kubernetes.io/projected/c2a7ac81-47a3-4c55-90bc-f19842c87e37-kube-api-access-r9t68\") pod \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.608267 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-config-data\") pod \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.608327 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-combined-ca-bundle\") pod \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.608420 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2a7ac81-47a3-4c55-90bc-f19842c87e37-logs\") pod \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\" (UID: \"c2a7ac81-47a3-4c55-90bc-f19842c87e37\") " Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.608897 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2a7ac81-47a3-4c55-90bc-f19842c87e37-logs" (OuterVolumeSpecName: "logs") pod "c2a7ac81-47a3-4c55-90bc-f19842c87e37" (UID: "c2a7ac81-47a3-4c55-90bc-f19842c87e37"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.612638 4931 scope.go:117] "RemoveContainer" containerID="88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b" Nov 29 00:22:17 crc kubenswrapper[4931]: E1129 00:22:17.613201 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b\": container with ID starting with 88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b not found: ID does not exist" containerID="88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.613224 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b"} err="failed to get container status \"88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b\": rpc error: code = NotFound desc = could not find container \"88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b\": container with ID starting with 88462ff79bc07e6731b9ba35d4a098e4229b064eca9b5e32bb8f39fce640387b not found: ID does not exist" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.613242 4931 scope.go:117] "RemoveContainer" containerID="65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff" Nov 29 00:22:17 crc kubenswrapper[4931]: E1129 00:22:17.613491 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff\": container with ID starting with 65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff not found: ID does not exist" containerID="65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.613509 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff"} err="failed to get container status \"65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff\": rpc error: code = NotFound desc = could not find container \"65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff\": container with ID starting with 65b55cae4e916e9ee9933fceef9d55e928d8d4ec69da10de1c659a24afbe8aff not found: ID does not exist" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.616443 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2a7ac81-47a3-4c55-90bc-f19842c87e37-kube-api-access-r9t68" (OuterVolumeSpecName: "kube-api-access-r9t68") pod "c2a7ac81-47a3-4c55-90bc-f19842c87e37" (UID: "c2a7ac81-47a3-4c55-90bc-f19842c87e37"). InnerVolumeSpecName "kube-api-access-r9t68". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.634174 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-config-data" (OuterVolumeSpecName: "config-data") pod "c2a7ac81-47a3-4c55-90bc-f19842c87e37" (UID: "c2a7ac81-47a3-4c55-90bc-f19842c87e37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.639860 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2a7ac81-47a3-4c55-90bc-f19842c87e37" (UID: "c2a7ac81-47a3-4c55-90bc-f19842c87e37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.695908 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c2a7ac81-47a3-4c55-90bc-f19842c87e37" (UID: "c2a7ac81-47a3-4c55-90bc-f19842c87e37"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.710313 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.710343 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2a7ac81-47a3-4c55-90bc-f19842c87e37-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.710353 4931 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.710362 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9t68\" (UniqueName: \"kubernetes.io/projected/c2a7ac81-47a3-4c55-90bc-f19842c87e37-kube-api-access-r9t68\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.710371 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2a7ac81-47a3-4c55-90bc-f19842c87e37-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.897739 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.908936 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.931828 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:22:17 crc kubenswrapper[4931]: E1129 00:22:17.932204 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerName="nova-metadata-log" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.932221 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerName="nova-metadata-log" Nov 29 00:22:17 crc kubenswrapper[4931]: E1129 00:22:17.932236 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerName="nova-metadata-metadata" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.932242 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerName="nova-metadata-metadata" Nov 29 00:22:17 crc kubenswrapper[4931]: E1129 00:22:17.932255 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="149e0e4c-6506-46d0-a6a4-965d8966c71e" containerName="dnsmasq-dns" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.932262 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="149e0e4c-6506-46d0-a6a4-965d8966c71e" containerName="dnsmasq-dns" Nov 29 00:22:17 crc kubenswrapper[4931]: E1129 00:22:17.932272 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f619a0b-81e1-4881-abaf-643f69860a61" containerName="nova-manage" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.932278 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f619a0b-81e1-4881-abaf-643f69860a61" containerName="nova-manage" Nov 29 00:22:17 crc kubenswrapper[4931]: E1129 00:22:17.932322 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="149e0e4c-6506-46d0-a6a4-965d8966c71e" containerName="init" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.932328 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="149e0e4c-6506-46d0-a6a4-965d8966c71e" containerName="init" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.932493 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerName="nova-metadata-log" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.932505 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="149e0e4c-6506-46d0-a6a4-965d8966c71e" containerName="dnsmasq-dns" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.932513 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" containerName="nova-metadata-metadata" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.932527 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f619a0b-81e1-4881-abaf-643f69860a61" containerName="nova-manage" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.933531 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.936394 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.936615 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 29 00:22:17 crc kubenswrapper[4931]: I1129 00:22:17.953779 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.014803 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6440833b-24fe-4da5-afae-6111fc7e760e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.014997 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6440833b-24fe-4da5-afae-6111fc7e760e-config-data\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.015092 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6440833b-24fe-4da5-afae-6111fc7e760e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.015149 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87vl7\" (UniqueName: \"kubernetes.io/projected/6440833b-24fe-4da5-afae-6111fc7e760e-kube-api-access-87vl7\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.015218 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6440833b-24fe-4da5-afae-6111fc7e760e-logs\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.116713 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6440833b-24fe-4da5-afae-6111fc7e760e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.116821 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87vl7\" (UniqueName: \"kubernetes.io/projected/6440833b-24fe-4da5-afae-6111fc7e760e-kube-api-access-87vl7\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.116924 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6440833b-24fe-4da5-afae-6111fc7e760e-logs\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.116972 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6440833b-24fe-4da5-afae-6111fc7e760e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.116997 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6440833b-24fe-4da5-afae-6111fc7e760e-config-data\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.117622 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6440833b-24fe-4da5-afae-6111fc7e760e-logs\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.121580 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6440833b-24fe-4da5-afae-6111fc7e760e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.123239 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6440833b-24fe-4da5-afae-6111fc7e760e-config-data\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.129743 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6440833b-24fe-4da5-afae-6111fc7e760e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.135785 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87vl7\" (UniqueName: \"kubernetes.io/projected/6440833b-24fe-4da5-afae-6111fc7e760e-kube-api-access-87vl7\") pod \"nova-metadata-0\" (UID: \"6440833b-24fe-4da5-afae-6111fc7e760e\") " pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.251261 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.576721 4931 generic.go:334] "Generic (PLEG): container finished" podID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" containerID="a476459aedbdc5c2cf64242efd1af0a97196ac1c973814ab7881d62669e67140" exitCode=0 Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.576762 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8a0b9809-f5c3-4426-92a6-7cff2a78b936","Type":"ContainerDied","Data":"a476459aedbdc5c2cf64242efd1af0a97196ac1c973814ab7881d62669e67140"} Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.695555 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.727121 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.830613 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-config-data\") pod \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.830979 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-public-tls-certs\") pod \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.831020 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-internal-tls-certs\") pod \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.831061 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cj25\" (UniqueName: \"kubernetes.io/projected/8a0b9809-f5c3-4426-92a6-7cff2a78b936-kube-api-access-5cj25\") pod \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.831212 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a0b9809-f5c3-4426-92a6-7cff2a78b936-logs\") pod \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.831238 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-combined-ca-bundle\") pod \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\" (UID: \"8a0b9809-f5c3-4426-92a6-7cff2a78b936\") " Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.833045 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a0b9809-f5c3-4426-92a6-7cff2a78b936-logs" (OuterVolumeSpecName: "logs") pod "8a0b9809-f5c3-4426-92a6-7cff2a78b936" (UID: "8a0b9809-f5c3-4426-92a6-7cff2a78b936"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.853872 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a0b9809-f5c3-4426-92a6-7cff2a78b936-kube-api-access-5cj25" (OuterVolumeSpecName: "kube-api-access-5cj25") pod "8a0b9809-f5c3-4426-92a6-7cff2a78b936" (UID: "8a0b9809-f5c3-4426-92a6-7cff2a78b936"). InnerVolumeSpecName "kube-api-access-5cj25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.874355 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-config-data" (OuterVolumeSpecName: "config-data") pod "8a0b9809-f5c3-4426-92a6-7cff2a78b936" (UID: "8a0b9809-f5c3-4426-92a6-7cff2a78b936"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.874772 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a0b9809-f5c3-4426-92a6-7cff2a78b936" (UID: "8a0b9809-f5c3-4426-92a6-7cff2a78b936"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.909699 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8a0b9809-f5c3-4426-92a6-7cff2a78b936" (UID: "8a0b9809-f5c3-4426-92a6-7cff2a78b936"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.914032 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8a0b9809-f5c3-4426-92a6-7cff2a78b936" (UID: "8a0b9809-f5c3-4426-92a6-7cff2a78b936"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.935745 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.935766 4931 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.935776 4931 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.935785 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cj25\" (UniqueName: \"kubernetes.io/projected/8a0b9809-f5c3-4426-92a6-7cff2a78b936-kube-api-access-5cj25\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.935794 4931 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a0b9809-f5c3-4426-92a6-7cff2a78b936-logs\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:18 crc kubenswrapper[4931]: I1129 00:22:18.935802 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a0b9809-f5c3-4426-92a6-7cff2a78b936-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.221799 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2a7ac81-47a3-4c55-90bc-f19842c87e37" path="/var/lib/kubelet/pods/c2a7ac81-47a3-4c55-90bc-f19842c87e37/volumes" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.573621 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.600347 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6440833b-24fe-4da5-afae-6111fc7e760e","Type":"ContainerStarted","Data":"efe5f9f8c4bdfcad0507d604741663c39da23badd162c93b5c46bc8f645cee95"} Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.600396 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6440833b-24fe-4da5-afae-6111fc7e760e","Type":"ContainerStarted","Data":"d71e5c00ba3b948e54e0f8cd22cae04facab5e1618ca72ba3b9d2ce31840ca1d"} Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.600409 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6440833b-24fe-4da5-afae-6111fc7e760e","Type":"ContainerStarted","Data":"433bcf18db71054e62320f7021d5a2e64764fcf5f55335cfae689b662ba74c55"} Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.615881 4931 generic.go:334] "Generic (PLEG): container finished" podID="d62bbf08-1341-44a7-b3d2-d85b0490f6f3" containerID="01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac" exitCode=0 Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.615965 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d62bbf08-1341-44a7-b3d2-d85b0490f6f3","Type":"ContainerDied","Data":"01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac"} Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.615992 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d62bbf08-1341-44a7-b3d2-d85b0490f6f3","Type":"ContainerDied","Data":"759b79ed53cd355b72261053c0ef3918b19147755c3d0b351a346c729c78722f"} Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.616012 4931 scope.go:117] "RemoveContainer" containerID="01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.616234 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.628999 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.628980248 podStartE2EDuration="2.628980248s" podCreationTimestamp="2025-11-29 00:22:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:22:19.6176541 +0000 UTC m=+1352.779547342" watchObservedRunningTime="2025-11-29 00:22:19.628980248 +0000 UTC m=+1352.790873480" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.632757 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8a0b9809-f5c3-4426-92a6-7cff2a78b936","Type":"ContainerDied","Data":"699342d05d5a38b3469f6402e8ab309286b541f1ea56471850248dea17ad93d4"} Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.633525 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.648413 4931 scope.go:117] "RemoveContainer" containerID="01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac" Nov 29 00:22:19 crc kubenswrapper[4931]: E1129 00:22:19.650360 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac\": container with ID starting with 01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac not found: ID does not exist" containerID="01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.650415 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac"} err="failed to get container status \"01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac\": rpc error: code = NotFound desc = could not find container \"01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac\": container with ID starting with 01a0851cf9d5da9faca5a2c4c57b8ca2639e77d205addd17aeaa00ae469a9cac not found: ID does not exist" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.650448 4931 scope.go:117] "RemoveContainer" containerID="a476459aedbdc5c2cf64242efd1af0a97196ac1c973814ab7881d62669e67140" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.684207 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.693621 4931 scope.go:117] "RemoveContainer" containerID="3d283e14a08e74b7385bdb154bac500e9b7449a59333d8c5669a009b9535ff67" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.710007 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.716123 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 29 00:22:19 crc kubenswrapper[4931]: E1129 00:22:19.716660 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d62bbf08-1341-44a7-b3d2-d85b0490f6f3" containerName="nova-scheduler-scheduler" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.716678 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="d62bbf08-1341-44a7-b3d2-d85b0490f6f3" containerName="nova-scheduler-scheduler" Nov 29 00:22:19 crc kubenswrapper[4931]: E1129 00:22:19.716748 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" containerName="nova-api-api" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.716757 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" containerName="nova-api-api" Nov 29 00:22:19 crc kubenswrapper[4931]: E1129 00:22:19.716776 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" containerName="nova-api-log" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.716783 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" containerName="nova-api-log" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.717058 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" containerName="nova-api-api" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.717096 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="d62bbf08-1341-44a7-b3d2-d85b0490f6f3" containerName="nova-scheduler-scheduler" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.717109 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" containerName="nova-api-log" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.718599 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.722082 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.722158 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.722179 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.729587 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.753894 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-config-data\") pod \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\" (UID: \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\") " Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.754084 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdk7w\" (UniqueName: \"kubernetes.io/projected/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-kube-api-access-gdk7w\") pod \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\" (UID: \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\") " Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.754128 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-combined-ca-bundle\") pod \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\" (UID: \"d62bbf08-1341-44a7-b3d2-d85b0490f6f3\") " Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.760322 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-kube-api-access-gdk7w" (OuterVolumeSpecName: "kube-api-access-gdk7w") pod "d62bbf08-1341-44a7-b3d2-d85b0490f6f3" (UID: "d62bbf08-1341-44a7-b3d2-d85b0490f6f3"). InnerVolumeSpecName "kube-api-access-gdk7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.788192 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d62bbf08-1341-44a7-b3d2-d85b0490f6f3" (UID: "d62bbf08-1341-44a7-b3d2-d85b0490f6f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.791655 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-config-data" (OuterVolumeSpecName: "config-data") pod "d62bbf08-1341-44a7-b3d2-d85b0490f6f3" (UID: "d62bbf08-1341-44a7-b3d2-d85b0490f6f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.856858 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/611072c6-340e-49fc-870e-8a1e56f2cc90-public-tls-certs\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.857190 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg7cm\" (UniqueName: \"kubernetes.io/projected/611072c6-340e-49fc-870e-8a1e56f2cc90-kube-api-access-wg7cm\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.857375 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/611072c6-340e-49fc-870e-8a1e56f2cc90-logs\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.857551 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611072c6-340e-49fc-870e-8a1e56f2cc90-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.857684 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/611072c6-340e-49fc-870e-8a1e56f2cc90-internal-tls-certs\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.857850 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611072c6-340e-49fc-870e-8a1e56f2cc90-config-data\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.858048 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.858148 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdk7w\" (UniqueName: \"kubernetes.io/projected/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-kube-api-access-gdk7w\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.858261 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d62bbf08-1341-44a7-b3d2-d85b0490f6f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.961873 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611072c6-340e-49fc-870e-8a1e56f2cc90-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.961987 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/611072c6-340e-49fc-870e-8a1e56f2cc90-internal-tls-certs\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.962093 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611072c6-340e-49fc-870e-8a1e56f2cc90-config-data\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.962183 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/611072c6-340e-49fc-870e-8a1e56f2cc90-public-tls-certs\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.962234 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg7cm\" (UniqueName: \"kubernetes.io/projected/611072c6-340e-49fc-870e-8a1e56f2cc90-kube-api-access-wg7cm\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.962329 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/611072c6-340e-49fc-870e-8a1e56f2cc90-logs\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.962723 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.963068 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/611072c6-340e-49fc-870e-8a1e56f2cc90-logs\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.966545 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611072c6-340e-49fc-870e-8a1e56f2cc90-config-data\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.966741 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/611072c6-340e-49fc-870e-8a1e56f2cc90-internal-tls-certs\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.967512 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611072c6-340e-49fc-870e-8a1e56f2cc90-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.972874 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/611072c6-340e-49fc-870e-8a1e56f2cc90-public-tls-certs\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.972939 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.977954 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg7cm\" (UniqueName: \"kubernetes.io/projected/611072c6-340e-49fc-870e-8a1e56f2cc90-kube-api-access-wg7cm\") pod \"nova-api-0\" (UID: \"611072c6-340e-49fc-870e-8a1e56f2cc90\") " pod="openstack/nova-api-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.984741 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.986215 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.988603 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 29 00:22:19 crc kubenswrapper[4931]: I1129 00:22:19.993174 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.040096 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.166463 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2-config-data\") pod \"nova-scheduler-0\" (UID: \"f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2\") " pod="openstack/nova-scheduler-0" Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.166851 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4cpt\" (UniqueName: \"kubernetes.io/projected/f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2-kube-api-access-l4cpt\") pod \"nova-scheduler-0\" (UID: \"f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2\") " pod="openstack/nova-scheduler-0" Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.166887 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2\") " pod="openstack/nova-scheduler-0" Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.273126 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2-config-data\") pod \"nova-scheduler-0\" (UID: \"f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2\") " pod="openstack/nova-scheduler-0" Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.273196 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4cpt\" (UniqueName: \"kubernetes.io/projected/f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2-kube-api-access-l4cpt\") pod \"nova-scheduler-0\" (UID: \"f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2\") " pod="openstack/nova-scheduler-0" Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.273253 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2\") " pod="openstack/nova-scheduler-0" Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.281654 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2\") " pod="openstack/nova-scheduler-0" Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.283458 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2-config-data\") pod \"nova-scheduler-0\" (UID: \"f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2\") " pod="openstack/nova-scheduler-0" Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.300886 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4cpt\" (UniqueName: \"kubernetes.io/projected/f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2-kube-api-access-l4cpt\") pod \"nova-scheduler-0\" (UID: \"f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2\") " pod="openstack/nova-scheduler-0" Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.317257 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.485158 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.649067 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"611072c6-340e-49fc-870e-8a1e56f2cc90","Type":"ContainerStarted","Data":"873d8addc9d38a0259670fd1c36f0fd903834b061795adf374685809eb756197"} Nov 29 00:22:20 crc kubenswrapper[4931]: I1129 00:22:20.888201 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 29 00:22:21 crc kubenswrapper[4931]: I1129 00:22:21.228912 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a0b9809-f5c3-4426-92a6-7cff2a78b936" path="/var/lib/kubelet/pods/8a0b9809-f5c3-4426-92a6-7cff2a78b936/volumes" Nov 29 00:22:21 crc kubenswrapper[4931]: I1129 00:22:21.229663 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d62bbf08-1341-44a7-b3d2-d85b0490f6f3" path="/var/lib/kubelet/pods/d62bbf08-1341-44a7-b3d2-d85b0490f6f3/volumes" Nov 29 00:22:21 crc kubenswrapper[4931]: I1129 00:22:21.660420 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2","Type":"ContainerStarted","Data":"0ab37f104fa926ba06f2c6404f751d7076ba8a30d66e5cdfae6eecb574a5b719"} Nov 29 00:22:21 crc kubenswrapper[4931]: I1129 00:22:21.660744 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2","Type":"ContainerStarted","Data":"02aefc63ded6cad27989064f31e9b64d3dc895633708b2b26b3c0ce19df24498"} Nov 29 00:22:21 crc kubenswrapper[4931]: I1129 00:22:21.663784 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"611072c6-340e-49fc-870e-8a1e56f2cc90","Type":"ContainerStarted","Data":"f379709bcd026e8c572be8d2473d6bd41b9f06a72fc538f48b2cb67c3e0cc254"} Nov 29 00:22:21 crc kubenswrapper[4931]: I1129 00:22:21.663841 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"611072c6-340e-49fc-870e-8a1e56f2cc90","Type":"ContainerStarted","Data":"d8f201cd63cf1fcb06cbca1fadfe05ff2114316ddec732bc987a2a64abf64a04"} Nov 29 00:22:21 crc kubenswrapper[4931]: I1129 00:22:21.714944 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.7149272509999998 podStartE2EDuration="2.714927251s" podCreationTimestamp="2025-11-29 00:22:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:22:21.685205911 +0000 UTC m=+1354.847099143" watchObservedRunningTime="2025-11-29 00:22:21.714927251 +0000 UTC m=+1354.876820483" Nov 29 00:22:21 crc kubenswrapper[4931]: I1129 00:22:21.717147 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.717137966 podStartE2EDuration="2.717137966s" podCreationTimestamp="2025-11-29 00:22:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:22:21.702315116 +0000 UTC m=+1354.864208368" watchObservedRunningTime="2025-11-29 00:22:21.717137966 +0000 UTC m=+1354.879031198" Nov 29 00:22:23 crc kubenswrapper[4931]: I1129 00:22:23.252259 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 29 00:22:23 crc kubenswrapper[4931]: I1129 00:22:23.252317 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 29 00:22:25 crc kubenswrapper[4931]: I1129 00:22:25.318442 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 29 00:22:28 crc kubenswrapper[4931]: I1129 00:22:28.251388 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 29 00:22:28 crc kubenswrapper[4931]: I1129 00:22:28.251952 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 29 00:22:29 crc kubenswrapper[4931]: I1129 00:22:29.266089 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6440833b-24fe-4da5-afae-6111fc7e760e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 29 00:22:29 crc kubenswrapper[4931]: I1129 00:22:29.266113 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6440833b-24fe-4da5-afae-6111fc7e760e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 00:22:30 crc kubenswrapper[4931]: I1129 00:22:30.040636 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 00:22:30 crc kubenswrapper[4931]: I1129 00:22:30.041013 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 29 00:22:30 crc kubenswrapper[4931]: I1129 00:22:30.317869 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 29 00:22:30 crc kubenswrapper[4931]: I1129 00:22:30.375002 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 29 00:22:30 crc kubenswrapper[4931]: I1129 00:22:30.819414 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 29 00:22:30 crc kubenswrapper[4931]: I1129 00:22:30.876234 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 29 00:22:31 crc kubenswrapper[4931]: I1129 00:22:31.061643 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="611072c6-340e-49fc-870e-8a1e56f2cc90" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 00:22:31 crc kubenswrapper[4931]: I1129 00:22:31.061663 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="611072c6-340e-49fc-870e-8a1e56f2cc90" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 29 00:22:36 crc kubenswrapper[4931]: I1129 00:22:36.975517 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6jfhm"] Nov 29 00:22:36 crc kubenswrapper[4931]: I1129 00:22:36.978341 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:37 crc kubenswrapper[4931]: I1129 00:22:37.007000 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6jfhm"] Nov 29 00:22:37 crc kubenswrapper[4931]: I1129 00:22:37.120791 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6aa684e-fed9-4f47-a2c5-165a35152b03-utilities\") pod \"redhat-operators-6jfhm\" (UID: \"c6aa684e-fed9-4f47-a2c5-165a35152b03\") " pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:37 crc kubenswrapper[4931]: I1129 00:22:37.121092 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6aa684e-fed9-4f47-a2c5-165a35152b03-catalog-content\") pod \"redhat-operators-6jfhm\" (UID: \"c6aa684e-fed9-4f47-a2c5-165a35152b03\") " pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:37 crc kubenswrapper[4931]: I1129 00:22:37.121216 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhxpj\" (UniqueName: \"kubernetes.io/projected/c6aa684e-fed9-4f47-a2c5-165a35152b03-kube-api-access-xhxpj\") pod \"redhat-operators-6jfhm\" (UID: \"c6aa684e-fed9-4f47-a2c5-165a35152b03\") " pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:37 crc kubenswrapper[4931]: I1129 00:22:37.222520 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhxpj\" (UniqueName: \"kubernetes.io/projected/c6aa684e-fed9-4f47-a2c5-165a35152b03-kube-api-access-xhxpj\") pod \"redhat-operators-6jfhm\" (UID: \"c6aa684e-fed9-4f47-a2c5-165a35152b03\") " pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:37 crc kubenswrapper[4931]: I1129 00:22:37.222613 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6aa684e-fed9-4f47-a2c5-165a35152b03-utilities\") pod \"redhat-operators-6jfhm\" (UID: \"c6aa684e-fed9-4f47-a2c5-165a35152b03\") " pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:37 crc kubenswrapper[4931]: I1129 00:22:37.222638 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6aa684e-fed9-4f47-a2c5-165a35152b03-catalog-content\") pod \"redhat-operators-6jfhm\" (UID: \"c6aa684e-fed9-4f47-a2c5-165a35152b03\") " pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:37 crc kubenswrapper[4931]: I1129 00:22:37.223067 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6aa684e-fed9-4f47-a2c5-165a35152b03-catalog-content\") pod \"redhat-operators-6jfhm\" (UID: \"c6aa684e-fed9-4f47-a2c5-165a35152b03\") " pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:37 crc kubenswrapper[4931]: I1129 00:22:37.223657 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6aa684e-fed9-4f47-a2c5-165a35152b03-utilities\") pod \"redhat-operators-6jfhm\" (UID: \"c6aa684e-fed9-4f47-a2c5-165a35152b03\") " pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:37 crc kubenswrapper[4931]: I1129 00:22:37.251348 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhxpj\" (UniqueName: \"kubernetes.io/projected/c6aa684e-fed9-4f47-a2c5-165a35152b03-kube-api-access-xhxpj\") pod \"redhat-operators-6jfhm\" (UID: \"c6aa684e-fed9-4f47-a2c5-165a35152b03\") " pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:37 crc kubenswrapper[4931]: I1129 00:22:37.312130 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:37 crc kubenswrapper[4931]: W1129 00:22:37.835336 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6aa684e_fed9_4f47_a2c5_165a35152b03.slice/crio-9fdfa6c384de800024d2b09b80ed14609344f1ee3f36194fde32fc7ac45b9a40 WatchSource:0}: Error finding container 9fdfa6c384de800024d2b09b80ed14609344f1ee3f36194fde32fc7ac45b9a40: Status 404 returned error can't find the container with id 9fdfa6c384de800024d2b09b80ed14609344f1ee3f36194fde32fc7ac45b9a40 Nov 29 00:22:37 crc kubenswrapper[4931]: I1129 00:22:37.850606 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6jfhm"] Nov 29 00:22:38 crc kubenswrapper[4931]: I1129 00:22:38.257410 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 29 00:22:38 crc kubenswrapper[4931]: I1129 00:22:38.259507 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 29 00:22:38 crc kubenswrapper[4931]: I1129 00:22:38.262362 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 29 00:22:38 crc kubenswrapper[4931]: I1129 00:22:38.848929 4931 generic.go:334] "Generic (PLEG): container finished" podID="c6aa684e-fed9-4f47-a2c5-165a35152b03" containerID="869a960d5a422f36543ea7b0c491cd3802ff1e0ec9aea473cd3bc77b6cde9f35" exitCode=0 Nov 29 00:22:38 crc kubenswrapper[4931]: I1129 00:22:38.849137 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jfhm" event={"ID":"c6aa684e-fed9-4f47-a2c5-165a35152b03","Type":"ContainerDied","Data":"869a960d5a422f36543ea7b0c491cd3802ff1e0ec9aea473cd3bc77b6cde9f35"} Nov 29 00:22:38 crc kubenswrapper[4931]: I1129 00:22:38.849202 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jfhm" event={"ID":"c6aa684e-fed9-4f47-a2c5-165a35152b03","Type":"ContainerStarted","Data":"9fdfa6c384de800024d2b09b80ed14609344f1ee3f36194fde32fc7ac45b9a40"} Nov 29 00:22:38 crc kubenswrapper[4931]: I1129 00:22:38.862244 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 29 00:22:40 crc kubenswrapper[4931]: I1129 00:22:40.049738 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 29 00:22:40 crc kubenswrapper[4931]: I1129 00:22:40.051391 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 29 00:22:40 crc kubenswrapper[4931]: I1129 00:22:40.051414 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 29 00:22:40 crc kubenswrapper[4931]: I1129 00:22:40.051427 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 29 00:22:40 crc kubenswrapper[4931]: I1129 00:22:40.057547 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 29 00:22:40 crc kubenswrapper[4931]: I1129 00:22:40.060790 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 29 00:22:40 crc kubenswrapper[4931]: I1129 00:22:40.874479 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jfhm" event={"ID":"c6aa684e-fed9-4f47-a2c5-165a35152b03","Type":"ContainerStarted","Data":"1a38119aac3e046ddcff1abef918313256c727bbfe1276eef76ed366c0bc02b7"} Nov 29 00:22:42 crc kubenswrapper[4931]: I1129 00:22:42.898301 4931 generic.go:334] "Generic (PLEG): container finished" podID="c6aa684e-fed9-4f47-a2c5-165a35152b03" containerID="1a38119aac3e046ddcff1abef918313256c727bbfe1276eef76ed366c0bc02b7" exitCode=0 Nov 29 00:22:42 crc kubenswrapper[4931]: I1129 00:22:42.898405 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jfhm" event={"ID":"c6aa684e-fed9-4f47-a2c5-165a35152b03","Type":"ContainerDied","Data":"1a38119aac3e046ddcff1abef918313256c727bbfe1276eef76ed366c0bc02b7"} Nov 29 00:22:44 crc kubenswrapper[4931]: I1129 00:22:44.919169 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jfhm" event={"ID":"c6aa684e-fed9-4f47-a2c5-165a35152b03","Type":"ContainerStarted","Data":"a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811"} Nov 29 00:22:44 crc kubenswrapper[4931]: I1129 00:22:44.945558 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6jfhm" podStartSLOduration=4.013414073 podStartE2EDuration="8.945540001s" podCreationTimestamp="2025-11-29 00:22:36 +0000 UTC" firstStartedPulling="2025-11-29 00:22:38.851481173 +0000 UTC m=+1372.013374405" lastFinishedPulling="2025-11-29 00:22:43.783607081 +0000 UTC m=+1376.945500333" observedRunningTime="2025-11-29 00:22:44.938474837 +0000 UTC m=+1378.100368089" watchObservedRunningTime="2025-11-29 00:22:44.945540001 +0000 UTC m=+1378.107433223" Nov 29 00:22:47 crc kubenswrapper[4931]: I1129 00:22:47.313220 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:47 crc kubenswrapper[4931]: I1129 00:22:47.314735 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:48 crc kubenswrapper[4931]: I1129 00:22:48.359523 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6jfhm" podUID="c6aa684e-fed9-4f47-a2c5-165a35152b03" containerName="registry-server" probeResult="failure" output=< Nov 29 00:22:48 crc kubenswrapper[4931]: timeout: failed to connect service ":50051" within 1s Nov 29 00:22:48 crc kubenswrapper[4931]: > Nov 29 00:22:57 crc kubenswrapper[4931]: I1129 00:22:57.360162 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:22:57 crc kubenswrapper[4931]: I1129 00:22:57.410073 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:23:01 crc kubenswrapper[4931]: I1129 00:23:01.493583 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6jfhm"] Nov 29 00:23:01 crc kubenswrapper[4931]: I1129 00:23:01.494465 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6jfhm" podUID="c6aa684e-fed9-4f47-a2c5-165a35152b03" containerName="registry-server" containerID="cri-o://a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811" gracePeriod=2 Nov 29 00:23:01 crc kubenswrapper[4931]: I1129 00:23:01.963066 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.080838 4931 generic.go:334] "Generic (PLEG): container finished" podID="c6aa684e-fed9-4f47-a2c5-165a35152b03" containerID="a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811" exitCode=0 Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.080886 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jfhm" event={"ID":"c6aa684e-fed9-4f47-a2c5-165a35152b03","Type":"ContainerDied","Data":"a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811"} Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.080912 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6jfhm" event={"ID":"c6aa684e-fed9-4f47-a2c5-165a35152b03","Type":"ContainerDied","Data":"9fdfa6c384de800024d2b09b80ed14609344f1ee3f36194fde32fc7ac45b9a40"} Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.080928 4931 scope.go:117] "RemoveContainer" containerID="a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.080986 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6jfhm" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.116445 4931 scope.go:117] "RemoveContainer" containerID="1a38119aac3e046ddcff1abef918313256c727bbfe1276eef76ed366c0bc02b7" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.139575 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6aa684e-fed9-4f47-a2c5-165a35152b03-utilities\") pod \"c6aa684e-fed9-4f47-a2c5-165a35152b03\" (UID: \"c6aa684e-fed9-4f47-a2c5-165a35152b03\") " Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.139669 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6aa684e-fed9-4f47-a2c5-165a35152b03-catalog-content\") pod \"c6aa684e-fed9-4f47-a2c5-165a35152b03\" (UID: \"c6aa684e-fed9-4f47-a2c5-165a35152b03\") " Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.139801 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhxpj\" (UniqueName: \"kubernetes.io/projected/c6aa684e-fed9-4f47-a2c5-165a35152b03-kube-api-access-xhxpj\") pod \"c6aa684e-fed9-4f47-a2c5-165a35152b03\" (UID: \"c6aa684e-fed9-4f47-a2c5-165a35152b03\") " Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.140797 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6aa684e-fed9-4f47-a2c5-165a35152b03-utilities" (OuterVolumeSpecName: "utilities") pod "c6aa684e-fed9-4f47-a2c5-165a35152b03" (UID: "c6aa684e-fed9-4f47-a2c5-165a35152b03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.148334 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6aa684e-fed9-4f47-a2c5-165a35152b03-kube-api-access-xhxpj" (OuterVolumeSpecName: "kube-api-access-xhxpj") pod "c6aa684e-fed9-4f47-a2c5-165a35152b03" (UID: "c6aa684e-fed9-4f47-a2c5-165a35152b03"). InnerVolumeSpecName "kube-api-access-xhxpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.163361 4931 scope.go:117] "RemoveContainer" containerID="869a960d5a422f36543ea7b0c491cd3802ff1e0ec9aea473cd3bc77b6cde9f35" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.224788 4931 scope.go:117] "RemoveContainer" containerID="a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811" Nov 29 00:23:02 crc kubenswrapper[4931]: E1129 00:23:02.225246 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811\": container with ID starting with a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811 not found: ID does not exist" containerID="a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.225288 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811"} err="failed to get container status \"a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811\": rpc error: code = NotFound desc = could not find container \"a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811\": container with ID starting with a475e19c815fe4b25889540fd0d304faf9e70b39d32a97215c545f4b57849811 not found: ID does not exist" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.225329 4931 scope.go:117] "RemoveContainer" containerID="1a38119aac3e046ddcff1abef918313256c727bbfe1276eef76ed366c0bc02b7" Nov 29 00:23:02 crc kubenswrapper[4931]: E1129 00:23:02.225620 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a38119aac3e046ddcff1abef918313256c727bbfe1276eef76ed366c0bc02b7\": container with ID starting with 1a38119aac3e046ddcff1abef918313256c727bbfe1276eef76ed366c0bc02b7 not found: ID does not exist" containerID="1a38119aac3e046ddcff1abef918313256c727bbfe1276eef76ed366c0bc02b7" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.225658 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a38119aac3e046ddcff1abef918313256c727bbfe1276eef76ed366c0bc02b7"} err="failed to get container status \"1a38119aac3e046ddcff1abef918313256c727bbfe1276eef76ed366c0bc02b7\": rpc error: code = NotFound desc = could not find container \"1a38119aac3e046ddcff1abef918313256c727bbfe1276eef76ed366c0bc02b7\": container with ID starting with 1a38119aac3e046ddcff1abef918313256c727bbfe1276eef76ed366c0bc02b7 not found: ID does not exist" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.225691 4931 scope.go:117] "RemoveContainer" containerID="869a960d5a422f36543ea7b0c491cd3802ff1e0ec9aea473cd3bc77b6cde9f35" Nov 29 00:23:02 crc kubenswrapper[4931]: E1129 00:23:02.225919 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"869a960d5a422f36543ea7b0c491cd3802ff1e0ec9aea473cd3bc77b6cde9f35\": container with ID starting with 869a960d5a422f36543ea7b0c491cd3802ff1e0ec9aea473cd3bc77b6cde9f35 not found: ID does not exist" containerID="869a960d5a422f36543ea7b0c491cd3802ff1e0ec9aea473cd3bc77b6cde9f35" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.225961 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"869a960d5a422f36543ea7b0c491cd3802ff1e0ec9aea473cd3bc77b6cde9f35"} err="failed to get container status \"869a960d5a422f36543ea7b0c491cd3802ff1e0ec9aea473cd3bc77b6cde9f35\": rpc error: code = NotFound desc = could not find container \"869a960d5a422f36543ea7b0c491cd3802ff1e0ec9aea473cd3bc77b6cde9f35\": container with ID starting with 869a960d5a422f36543ea7b0c491cd3802ff1e0ec9aea473cd3bc77b6cde9f35 not found: ID does not exist" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.243126 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6aa684e-fed9-4f47-a2c5-165a35152b03-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.243170 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhxpj\" (UniqueName: \"kubernetes.io/projected/c6aa684e-fed9-4f47-a2c5-165a35152b03-kube-api-access-xhxpj\") on node \"crc\" DevicePath \"\"" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.258495 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6aa684e-fed9-4f47-a2c5-165a35152b03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6aa684e-fed9-4f47-a2c5-165a35152b03" (UID: "c6aa684e-fed9-4f47-a2c5-165a35152b03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.345206 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6aa684e-fed9-4f47-a2c5-165a35152b03-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.419320 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6jfhm"] Nov 29 00:23:02 crc kubenswrapper[4931]: I1129 00:23:02.427577 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6jfhm"] Nov 29 00:23:03 crc kubenswrapper[4931]: I1129 00:23:03.233076 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6aa684e-fed9-4f47-a2c5-165a35152b03" path="/var/lib/kubelet/pods/c6aa684e-fed9-4f47-a2c5-165a35152b03/volumes" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.320384 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9vq9r"] Nov 29 00:23:22 crc kubenswrapper[4931]: E1129 00:23:22.321280 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6aa684e-fed9-4f47-a2c5-165a35152b03" containerName="registry-server" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.321292 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6aa684e-fed9-4f47-a2c5-165a35152b03" containerName="registry-server" Nov 29 00:23:22 crc kubenswrapper[4931]: E1129 00:23:22.321304 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6aa684e-fed9-4f47-a2c5-165a35152b03" containerName="extract-utilities" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.321311 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6aa684e-fed9-4f47-a2c5-165a35152b03" containerName="extract-utilities" Nov 29 00:23:22 crc kubenswrapper[4931]: E1129 00:23:22.321318 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6aa684e-fed9-4f47-a2c5-165a35152b03" containerName="extract-content" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.321325 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6aa684e-fed9-4f47-a2c5-165a35152b03" containerName="extract-content" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.321504 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6aa684e-fed9-4f47-a2c5-165a35152b03" containerName="registry-server" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.323874 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.330565 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vq9r"] Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.336715 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlmjr\" (UniqueName: \"kubernetes.io/projected/28742c83-1a46-4d33-8f7c-77e6fcc254d8-kube-api-access-zlmjr\") pod \"redhat-marketplace-9vq9r\" (UID: \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\") " pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.336777 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28742c83-1a46-4d33-8f7c-77e6fcc254d8-catalog-content\") pod \"redhat-marketplace-9vq9r\" (UID: \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\") " pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.336854 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28742c83-1a46-4d33-8f7c-77e6fcc254d8-utilities\") pod \"redhat-marketplace-9vq9r\" (UID: \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\") " pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.438217 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28742c83-1a46-4d33-8f7c-77e6fcc254d8-utilities\") pod \"redhat-marketplace-9vq9r\" (UID: \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\") " pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.438431 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlmjr\" (UniqueName: \"kubernetes.io/projected/28742c83-1a46-4d33-8f7c-77e6fcc254d8-kube-api-access-zlmjr\") pod \"redhat-marketplace-9vq9r\" (UID: \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\") " pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.438518 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28742c83-1a46-4d33-8f7c-77e6fcc254d8-catalog-content\") pod \"redhat-marketplace-9vq9r\" (UID: \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\") " pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.438734 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28742c83-1a46-4d33-8f7c-77e6fcc254d8-utilities\") pod \"redhat-marketplace-9vq9r\" (UID: \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\") " pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.438995 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28742c83-1a46-4d33-8f7c-77e6fcc254d8-catalog-content\") pod \"redhat-marketplace-9vq9r\" (UID: \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\") " pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.472563 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlmjr\" (UniqueName: \"kubernetes.io/projected/28742c83-1a46-4d33-8f7c-77e6fcc254d8-kube-api-access-zlmjr\") pod \"redhat-marketplace-9vq9r\" (UID: \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\") " pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:22 crc kubenswrapper[4931]: I1129 00:23:22.652713 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:23 crc kubenswrapper[4931]: W1129 00:23:23.221269 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28742c83_1a46_4d33_8f7c_77e6fcc254d8.slice/crio-069fa2a13a17a6d53cedc12d93867be7ef79a176ceed9ac5a7b411f9a6cff1bc WatchSource:0}: Error finding container 069fa2a13a17a6d53cedc12d93867be7ef79a176ceed9ac5a7b411f9a6cff1bc: Status 404 returned error can't find the container with id 069fa2a13a17a6d53cedc12d93867be7ef79a176ceed9ac5a7b411f9a6cff1bc Nov 29 00:23:23 crc kubenswrapper[4931]: I1129 00:23:23.229007 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vq9r"] Nov 29 00:23:23 crc kubenswrapper[4931]: I1129 00:23:23.289504 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vq9r" event={"ID":"28742c83-1a46-4d33-8f7c-77e6fcc254d8","Type":"ContainerStarted","Data":"069fa2a13a17a6d53cedc12d93867be7ef79a176ceed9ac5a7b411f9a6cff1bc"} Nov 29 00:23:23 crc kubenswrapper[4931]: I1129 00:23:23.734797 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:23:23 crc kubenswrapper[4931]: I1129 00:23:23.735180 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:23:24 crc kubenswrapper[4931]: I1129 00:23:24.298711 4931 generic.go:334] "Generic (PLEG): container finished" podID="28742c83-1a46-4d33-8f7c-77e6fcc254d8" containerID="07a241a477c903595c6c22a3fcd156b30a89bf4f2f01ad6fca12e926adddd5cb" exitCode=0 Nov 29 00:23:24 crc kubenswrapper[4931]: I1129 00:23:24.298761 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vq9r" event={"ID":"28742c83-1a46-4d33-8f7c-77e6fcc254d8","Type":"ContainerDied","Data":"07a241a477c903595c6c22a3fcd156b30a89bf4f2f01ad6fca12e926adddd5cb"} Nov 29 00:23:25 crc kubenswrapper[4931]: I1129 00:23:25.310456 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vq9r" event={"ID":"28742c83-1a46-4d33-8f7c-77e6fcc254d8","Type":"ContainerStarted","Data":"926c400b5913401abee9d03649b183d91b18665abc2c9095120f8d4a5c6dfc0d"} Nov 29 00:23:26 crc kubenswrapper[4931]: I1129 00:23:26.328648 4931 generic.go:334] "Generic (PLEG): container finished" podID="28742c83-1a46-4d33-8f7c-77e6fcc254d8" containerID="926c400b5913401abee9d03649b183d91b18665abc2c9095120f8d4a5c6dfc0d" exitCode=0 Nov 29 00:23:26 crc kubenswrapper[4931]: I1129 00:23:26.328727 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vq9r" event={"ID":"28742c83-1a46-4d33-8f7c-77e6fcc254d8","Type":"ContainerDied","Data":"926c400b5913401abee9d03649b183d91b18665abc2c9095120f8d4a5c6dfc0d"} Nov 29 00:23:28 crc kubenswrapper[4931]: I1129 00:23:28.354147 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vq9r" event={"ID":"28742c83-1a46-4d33-8f7c-77e6fcc254d8","Type":"ContainerStarted","Data":"93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef"} Nov 29 00:23:28 crc kubenswrapper[4931]: I1129 00:23:28.379637 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9vq9r" podStartSLOduration=3.350913747 podStartE2EDuration="6.379621882s" podCreationTimestamp="2025-11-29 00:23:22 +0000 UTC" firstStartedPulling="2025-11-29 00:23:24.301063145 +0000 UTC m=+1417.462956377" lastFinishedPulling="2025-11-29 00:23:27.32977127 +0000 UTC m=+1420.491664512" observedRunningTime="2025-11-29 00:23:28.377246424 +0000 UTC m=+1421.539139676" watchObservedRunningTime="2025-11-29 00:23:28.379621882 +0000 UTC m=+1421.541515114" Nov 29 00:23:32 crc kubenswrapper[4931]: I1129 00:23:32.654276 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:32 crc kubenswrapper[4931]: I1129 00:23:32.654915 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:32 crc kubenswrapper[4931]: I1129 00:23:32.716706 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:33 crc kubenswrapper[4931]: I1129 00:23:33.466405 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:35 crc kubenswrapper[4931]: I1129 00:23:35.864142 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vq9r"] Nov 29 00:23:35 crc kubenswrapper[4931]: I1129 00:23:35.865000 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9vq9r" podUID="28742c83-1a46-4d33-8f7c-77e6fcc254d8" containerName="registry-server" containerID="cri-o://93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef" gracePeriod=2 Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.345603 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.442420 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlmjr\" (UniqueName: \"kubernetes.io/projected/28742c83-1a46-4d33-8f7c-77e6fcc254d8-kube-api-access-zlmjr\") pod \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\" (UID: \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\") " Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.442555 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28742c83-1a46-4d33-8f7c-77e6fcc254d8-utilities\") pod \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\" (UID: \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\") " Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.442639 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28742c83-1a46-4d33-8f7c-77e6fcc254d8-catalog-content\") pod \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\" (UID: \"28742c83-1a46-4d33-8f7c-77e6fcc254d8\") " Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.450127 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28742c83-1a46-4d33-8f7c-77e6fcc254d8-utilities" (OuterVolumeSpecName: "utilities") pod "28742c83-1a46-4d33-8f7c-77e6fcc254d8" (UID: "28742c83-1a46-4d33-8f7c-77e6fcc254d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.450946 4931 generic.go:334] "Generic (PLEG): container finished" podID="28742c83-1a46-4d33-8f7c-77e6fcc254d8" containerID="93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef" exitCode=0 Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.451003 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vq9r" event={"ID":"28742c83-1a46-4d33-8f7c-77e6fcc254d8","Type":"ContainerDied","Data":"93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef"} Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.451030 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vq9r" event={"ID":"28742c83-1a46-4d33-8f7c-77e6fcc254d8","Type":"ContainerDied","Data":"069fa2a13a17a6d53cedc12d93867be7ef79a176ceed9ac5a7b411f9a6cff1bc"} Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.451046 4931 scope.go:117] "RemoveContainer" containerID="93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.451061 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vq9r" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.453030 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28742c83-1a46-4d33-8f7c-77e6fcc254d8-kube-api-access-zlmjr" (OuterVolumeSpecName: "kube-api-access-zlmjr") pod "28742c83-1a46-4d33-8f7c-77e6fcc254d8" (UID: "28742c83-1a46-4d33-8f7c-77e6fcc254d8"). InnerVolumeSpecName "kube-api-access-zlmjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.466549 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28742c83-1a46-4d33-8f7c-77e6fcc254d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28742c83-1a46-4d33-8f7c-77e6fcc254d8" (UID: "28742c83-1a46-4d33-8f7c-77e6fcc254d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.512269 4931 scope.go:117] "RemoveContainer" containerID="926c400b5913401abee9d03649b183d91b18665abc2c9095120f8d4a5c6dfc0d" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.529246 4931 scope.go:117] "RemoveContainer" containerID="07a241a477c903595c6c22a3fcd156b30a89bf4f2f01ad6fca12e926adddd5cb" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.544919 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28742c83-1a46-4d33-8f7c-77e6fcc254d8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.545190 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlmjr\" (UniqueName: \"kubernetes.io/projected/28742c83-1a46-4d33-8f7c-77e6fcc254d8-kube-api-access-zlmjr\") on node \"crc\" DevicePath \"\"" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.545261 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28742c83-1a46-4d33-8f7c-77e6fcc254d8-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.567536 4931 scope.go:117] "RemoveContainer" containerID="93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef" Nov 29 00:23:36 crc kubenswrapper[4931]: E1129 00:23:36.567933 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef\": container with ID starting with 93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef not found: ID does not exist" containerID="93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.568150 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef"} err="failed to get container status \"93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef\": rpc error: code = NotFound desc = could not find container \"93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef\": container with ID starting with 93f53c7a268e97f26ba36e7b5a1cacbec286f03cf8e2b511a597a509c50c6eef not found: ID does not exist" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.568280 4931 scope.go:117] "RemoveContainer" containerID="926c400b5913401abee9d03649b183d91b18665abc2c9095120f8d4a5c6dfc0d" Nov 29 00:23:36 crc kubenswrapper[4931]: E1129 00:23:36.568684 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"926c400b5913401abee9d03649b183d91b18665abc2c9095120f8d4a5c6dfc0d\": container with ID starting with 926c400b5913401abee9d03649b183d91b18665abc2c9095120f8d4a5c6dfc0d not found: ID does not exist" containerID="926c400b5913401abee9d03649b183d91b18665abc2c9095120f8d4a5c6dfc0d" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.568744 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"926c400b5913401abee9d03649b183d91b18665abc2c9095120f8d4a5c6dfc0d"} err="failed to get container status \"926c400b5913401abee9d03649b183d91b18665abc2c9095120f8d4a5c6dfc0d\": rpc error: code = NotFound desc = could not find container \"926c400b5913401abee9d03649b183d91b18665abc2c9095120f8d4a5c6dfc0d\": container with ID starting with 926c400b5913401abee9d03649b183d91b18665abc2c9095120f8d4a5c6dfc0d not found: ID does not exist" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.568780 4931 scope.go:117] "RemoveContainer" containerID="07a241a477c903595c6c22a3fcd156b30a89bf4f2f01ad6fca12e926adddd5cb" Nov 29 00:23:36 crc kubenswrapper[4931]: E1129 00:23:36.569057 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07a241a477c903595c6c22a3fcd156b30a89bf4f2f01ad6fca12e926adddd5cb\": container with ID starting with 07a241a477c903595c6c22a3fcd156b30a89bf4f2f01ad6fca12e926adddd5cb not found: ID does not exist" containerID="07a241a477c903595c6c22a3fcd156b30a89bf4f2f01ad6fca12e926adddd5cb" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.569092 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07a241a477c903595c6c22a3fcd156b30a89bf4f2f01ad6fca12e926adddd5cb"} err="failed to get container status \"07a241a477c903595c6c22a3fcd156b30a89bf4f2f01ad6fca12e926adddd5cb\": rpc error: code = NotFound desc = could not find container \"07a241a477c903595c6c22a3fcd156b30a89bf4f2f01ad6fca12e926adddd5cb\": container with ID starting with 07a241a477c903595c6c22a3fcd156b30a89bf4f2f01ad6fca12e926adddd5cb not found: ID does not exist" Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.782446 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vq9r"] Nov 29 00:23:36 crc kubenswrapper[4931]: I1129 00:23:36.790785 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vq9r"] Nov 29 00:23:37 crc kubenswrapper[4931]: I1129 00:23:37.220681 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28742c83-1a46-4d33-8f7c-77e6fcc254d8" path="/var/lib/kubelet/pods/28742c83-1a46-4d33-8f7c-77e6fcc254d8/volumes" Nov 29 00:23:53 crc kubenswrapper[4931]: I1129 00:23:53.734959 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:23:53 crc kubenswrapper[4931]: I1129 00:23:53.735498 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:23:54 crc kubenswrapper[4931]: E1129 00:23:54.178633 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-storage-0" podUID="a720d879-8b09-452d-8371-9bf75f28cbde" Nov 29 00:23:54 crc kubenswrapper[4931]: I1129 00:23:54.644702 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 29 00:23:56 crc kubenswrapper[4931]: I1129 00:23:56.067676 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:23:56 crc kubenswrapper[4931]: E1129 00:23:56.067888 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:23:56 crc kubenswrapper[4931]: E1129 00:23:56.068066 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 29 00:23:56 crc kubenswrapper[4931]: E1129 00:23:56.068133 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift podName:a720d879-8b09-452d-8371-9bf75f28cbde nodeName:}" failed. No retries permitted until 2025-11-29 00:25:58.068114643 +0000 UTC m=+1571.230007895 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift") pod "swift-storage-0" (UID: "a720d879-8b09-452d-8371-9bf75f28cbde") : configmap "swift-ring-files" not found Nov 29 00:24:10 crc kubenswrapper[4931]: E1129 00:24:10.467065 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-proxy-764cb44577-7q67f" podUID="550e16f7-f101-4a91-8e2f-45af3430d6b9" Nov 29 00:24:10 crc kubenswrapper[4931]: I1129 00:24:10.844896 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:24:13 crc kubenswrapper[4931]: I1129 00:24:13.565265 4931 scope.go:117] "RemoveContainer" containerID="97a836d59e02d3efb3146f155216201d667c13ac2481ba56d3133411b26b3ae9" Nov 29 00:24:14 crc kubenswrapper[4931]: I1129 00:24:14.282040 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:24:14 crc kubenswrapper[4931]: E1129 00:24:14.282579 4931 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 29 00:24:14 crc kubenswrapper[4931]: E1129 00:24:14.282696 4931 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-764cb44577-7q67f: configmap "swift-ring-files" not found Nov 29 00:24:14 crc kubenswrapper[4931]: E1129 00:24:14.282857 4931 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift podName:550e16f7-f101-4a91-8e2f-45af3430d6b9 nodeName:}" failed. No retries permitted until 2025-11-29 00:26:16.282833371 +0000 UTC m=+1589.444726613 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift") pod "swift-proxy-764cb44577-7q67f" (UID: "550e16f7-f101-4a91-8e2f-45af3430d6b9") : configmap "swift-ring-files" not found Nov 29 00:24:23 crc kubenswrapper[4931]: I1129 00:24:23.735719 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:24:23 crc kubenswrapper[4931]: I1129 00:24:23.736368 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:24:23 crc kubenswrapper[4931]: I1129 00:24:23.736456 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:24:23 crc kubenswrapper[4931]: I1129 00:24:23.737594 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e130783d6b1a728f7dced4948fd5956d2a8b18dc3e74e1cf5f76c0244465eeba"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 00:24:23 crc kubenswrapper[4931]: I1129 00:24:23.737681 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://e130783d6b1a728f7dced4948fd5956d2a8b18dc3e74e1cf5f76c0244465eeba" gracePeriod=600 Nov 29 00:24:24 crc kubenswrapper[4931]: I1129 00:24:24.009385 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="e130783d6b1a728f7dced4948fd5956d2a8b18dc3e74e1cf5f76c0244465eeba" exitCode=0 Nov 29 00:24:24 crc kubenswrapper[4931]: I1129 00:24:24.009433 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"e130783d6b1a728f7dced4948fd5956d2a8b18dc3e74e1cf5f76c0244465eeba"} Nov 29 00:24:24 crc kubenswrapper[4931]: I1129 00:24:24.009951 4931 scope.go:117] "RemoveContainer" containerID="7ca6926ea091667ee0b6b53772fafdc3830c2a5a094ad559af9957d016e82d86" Nov 29 00:24:25 crc kubenswrapper[4931]: I1129 00:24:25.023379 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357"} Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.626883 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-jnxgv"] Nov 29 00:24:30 crc kubenswrapper[4931]: E1129 00:24:30.628003 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28742c83-1a46-4d33-8f7c-77e6fcc254d8" containerName="registry-server" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.628026 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="28742c83-1a46-4d33-8f7c-77e6fcc254d8" containerName="registry-server" Nov 29 00:24:30 crc kubenswrapper[4931]: E1129 00:24:30.628062 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28742c83-1a46-4d33-8f7c-77e6fcc254d8" containerName="extract-content" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.628072 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="28742c83-1a46-4d33-8f7c-77e6fcc254d8" containerName="extract-content" Nov 29 00:24:30 crc kubenswrapper[4931]: E1129 00:24:30.628091 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28742c83-1a46-4d33-8f7c-77e6fcc254d8" containerName="extract-utilities" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.628099 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="28742c83-1a46-4d33-8f7c-77e6fcc254d8" containerName="extract-utilities" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.628365 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="28742c83-1a46-4d33-8f7c-77e6fcc254d8" containerName="registry-server" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.629196 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.631370 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.633048 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.640474 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jnxgv"] Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.722088 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-dispersionconf\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.722139 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64db2499-e411-42b9-8e99-92e748a9ad5d-etc-swift\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.722191 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-combined-ca-bundle\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.722552 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-swiftconf\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.722614 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64db2499-e411-42b9-8e99-92e748a9ad5d-ring-data-devices\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.722673 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d5pr\" (UniqueName: \"kubernetes.io/projected/64db2499-e411-42b9-8e99-92e748a9ad5d-kube-api-access-4d5pr\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.722914 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64db2499-e411-42b9-8e99-92e748a9ad5d-scripts\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.824670 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-swiftconf\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.825145 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64db2499-e411-42b9-8e99-92e748a9ad5d-ring-data-devices\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.825348 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d5pr\" (UniqueName: \"kubernetes.io/projected/64db2499-e411-42b9-8e99-92e748a9ad5d-kube-api-access-4d5pr\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.825604 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64db2499-e411-42b9-8e99-92e748a9ad5d-scripts\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.825849 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-dispersionconf\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.826054 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64db2499-e411-42b9-8e99-92e748a9ad5d-etc-swift\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.826266 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-combined-ca-bundle\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.826386 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64db2499-e411-42b9-8e99-92e748a9ad5d-etc-swift\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.826436 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64db2499-e411-42b9-8e99-92e748a9ad5d-ring-data-devices\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.826714 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64db2499-e411-42b9-8e99-92e748a9ad5d-scripts\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.832921 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-combined-ca-bundle\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.833057 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-dispersionconf\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.833423 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-swiftconf\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:30 crc kubenswrapper[4931]: I1129 00:24:30.858674 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d5pr\" (UniqueName: \"kubernetes.io/projected/64db2499-e411-42b9-8e99-92e748a9ad5d-kube-api-access-4d5pr\") pod \"swift-ring-rebalance-jnxgv\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:31 crc kubenswrapper[4931]: I1129 00:24:31.004414 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-mmzf8" Nov 29 00:24:31 crc kubenswrapper[4931]: I1129 00:24:31.012211 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:31 crc kubenswrapper[4931]: I1129 00:24:31.506396 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jnxgv"] Nov 29 00:24:32 crc kubenswrapper[4931]: I1129 00:24:32.095354 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jnxgv" event={"ID":"64db2499-e411-42b9-8e99-92e748a9ad5d","Type":"ContainerStarted","Data":"5daf90bcc18ccd95710faf4195a81591317bb972dc95dcfb8222b137a5c4480b"} Nov 29 00:24:36 crc kubenswrapper[4931]: I1129 00:24:36.141079 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jnxgv" event={"ID":"64db2499-e411-42b9-8e99-92e748a9ad5d","Type":"ContainerStarted","Data":"d328bf192247d8acf1c78a8259d9a7515bfb0575469bfeb89bb58d12b67492c3"} Nov 29 00:24:36 crc kubenswrapper[4931]: I1129 00:24:36.167891 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-jnxgv" podStartSLOduration=2.026370473 podStartE2EDuration="6.167870089s" podCreationTimestamp="2025-11-29 00:24:30 +0000 UTC" firstStartedPulling="2025-11-29 00:24:31.508717319 +0000 UTC m=+1484.670610561" lastFinishedPulling="2025-11-29 00:24:35.650216945 +0000 UTC m=+1488.812110177" observedRunningTime="2025-11-29 00:24:36.159143048 +0000 UTC m=+1489.321036300" watchObservedRunningTime="2025-11-29 00:24:36.167870089 +0000 UTC m=+1489.329763341" Nov 29 00:24:45 crc kubenswrapper[4931]: I1129 00:24:45.231694 4931 generic.go:334] "Generic (PLEG): container finished" podID="64db2499-e411-42b9-8e99-92e748a9ad5d" containerID="d328bf192247d8acf1c78a8259d9a7515bfb0575469bfeb89bb58d12b67492c3" exitCode=0 Nov 29 00:24:45 crc kubenswrapper[4931]: I1129 00:24:45.231842 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jnxgv" event={"ID":"64db2499-e411-42b9-8e99-92e748a9ad5d","Type":"ContainerDied","Data":"d328bf192247d8acf1c78a8259d9a7515bfb0575469bfeb89bb58d12b67492c3"} Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.703730 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.839928 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d5pr\" (UniqueName: \"kubernetes.io/projected/64db2499-e411-42b9-8e99-92e748a9ad5d-kube-api-access-4d5pr\") pod \"64db2499-e411-42b9-8e99-92e748a9ad5d\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.840346 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64db2499-e411-42b9-8e99-92e748a9ad5d-ring-data-devices\") pod \"64db2499-e411-42b9-8e99-92e748a9ad5d\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.840405 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-combined-ca-bundle\") pod \"64db2499-e411-42b9-8e99-92e748a9ad5d\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.840449 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-dispersionconf\") pod \"64db2499-e411-42b9-8e99-92e748a9ad5d\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.840511 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64db2499-e411-42b9-8e99-92e748a9ad5d-etc-swift\") pod \"64db2499-e411-42b9-8e99-92e748a9ad5d\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.840572 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64db2499-e411-42b9-8e99-92e748a9ad5d-scripts\") pod \"64db2499-e411-42b9-8e99-92e748a9ad5d\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.840607 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-swiftconf\") pod \"64db2499-e411-42b9-8e99-92e748a9ad5d\" (UID: \"64db2499-e411-42b9-8e99-92e748a9ad5d\") " Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.840914 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64db2499-e411-42b9-8e99-92e748a9ad5d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "64db2499-e411-42b9-8e99-92e748a9ad5d" (UID: "64db2499-e411-42b9-8e99-92e748a9ad5d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.841499 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64db2499-e411-42b9-8e99-92e748a9ad5d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "64db2499-e411-42b9-8e99-92e748a9ad5d" (UID: "64db2499-e411-42b9-8e99-92e748a9ad5d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.841656 4931 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64db2499-e411-42b9-8e99-92e748a9ad5d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.841695 4931 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64db2499-e411-42b9-8e99-92e748a9ad5d-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.863075 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64db2499-e411-42b9-8e99-92e748a9ad5d-kube-api-access-4d5pr" (OuterVolumeSpecName: "kube-api-access-4d5pr") pod "64db2499-e411-42b9-8e99-92e748a9ad5d" (UID: "64db2499-e411-42b9-8e99-92e748a9ad5d"). InnerVolumeSpecName "kube-api-access-4d5pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.871241 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64db2499-e411-42b9-8e99-92e748a9ad5d-scripts" (OuterVolumeSpecName: "scripts") pod "64db2499-e411-42b9-8e99-92e748a9ad5d" (UID: "64db2499-e411-42b9-8e99-92e748a9ad5d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.873556 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64db2499-e411-42b9-8e99-92e748a9ad5d" (UID: "64db2499-e411-42b9-8e99-92e748a9ad5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.874114 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "64db2499-e411-42b9-8e99-92e748a9ad5d" (UID: "64db2499-e411-42b9-8e99-92e748a9ad5d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.900312 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "64db2499-e411-42b9-8e99-92e748a9ad5d" (UID: "64db2499-e411-42b9-8e99-92e748a9ad5d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.942476 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d5pr\" (UniqueName: \"kubernetes.io/projected/64db2499-e411-42b9-8e99-92e748a9ad5d-kube-api-access-4d5pr\") on node \"crc\" DevicePath \"\"" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.942508 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.942517 4931 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.942527 4931 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64db2499-e411-42b9-8e99-92e748a9ad5d-scripts\") on node \"crc\" DevicePath \"\"" Nov 29 00:24:46 crc kubenswrapper[4931]: I1129 00:24:46.942536 4931 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64db2499-e411-42b9-8e99-92e748a9ad5d-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 29 00:24:47 crc kubenswrapper[4931]: I1129 00:24:47.263211 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jnxgv" event={"ID":"64db2499-e411-42b9-8e99-92e748a9ad5d","Type":"ContainerDied","Data":"5daf90bcc18ccd95710faf4195a81591317bb972dc95dcfb8222b137a5c4480b"} Nov 29 00:24:47 crc kubenswrapper[4931]: I1129 00:24:47.263265 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5daf90bcc18ccd95710faf4195a81591317bb972dc95dcfb8222b137a5c4480b" Nov 29 00:24:47 crc kubenswrapper[4931]: I1129 00:24:47.263245 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jnxgv" Nov 29 00:24:53 crc kubenswrapper[4931]: E1129 00:24:53.618825 4931 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db2499_e411_42b9_8e99_92e748a9ad5d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db2499_e411_42b9_8e99_92e748a9ad5d.slice/crio-5daf90bcc18ccd95710faf4195a81591317bb972dc95dcfb8222b137a5c4480b\": RecentStats: unable to find data in memory cache]" Nov 29 00:25:03 crc kubenswrapper[4931]: E1129 00:25:03.915459 4931 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db2499_e411_42b9_8e99_92e748a9ad5d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db2499_e411_42b9_8e99_92e748a9ad5d.slice/crio-5daf90bcc18ccd95710faf4195a81591317bb972dc95dcfb8222b137a5c4480b\": RecentStats: unable to find data in memory cache]" Nov 29 00:25:13 crc kubenswrapper[4931]: I1129 00:25:13.657595 4931 scope.go:117] "RemoveContainer" containerID="fca4eb12fdb64a7cbd05f5848df54f0703027b9db8e4943267818e18d15a5ca0" Nov 29 00:25:14 crc kubenswrapper[4931]: E1129 00:25:14.199342 4931 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db2499_e411_42b9_8e99_92e748a9ad5d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db2499_e411_42b9_8e99_92e748a9ad5d.slice/crio-5daf90bcc18ccd95710faf4195a81591317bb972dc95dcfb8222b137a5c4480b\": RecentStats: unable to find data in memory cache]" Nov 29 00:25:24 crc kubenswrapper[4931]: E1129 00:25:24.473102 4931 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db2499_e411_42b9_8e99_92e748a9ad5d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db2499_e411_42b9_8e99_92e748a9ad5d.slice/crio-5daf90bcc18ccd95710faf4195a81591317bb972dc95dcfb8222b137a5c4480b\": RecentStats: unable to find data in memory cache]" Nov 29 00:25:34 crc kubenswrapper[4931]: E1129 00:25:34.736398 4931 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db2499_e411_42b9_8e99_92e748a9ad5d.slice/crio-5daf90bcc18ccd95710faf4195a81591317bb972dc95dcfb8222b137a5c4480b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db2499_e411_42b9_8e99_92e748a9ad5d.slice\": RecentStats: unable to find data in memory cache]" Nov 29 00:25:45 crc kubenswrapper[4931]: E1129 00:25:45.077111 4931 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db2499_e411_42b9_8e99_92e748a9ad5d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64db2499_e411_42b9_8e99_92e748a9ad5d.slice/crio-5daf90bcc18ccd95710faf4195a81591317bb972dc95dcfb8222b137a5c4480b\": RecentStats: unable to find data in memory cache]" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.693653 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cvwd7"] Nov 29 00:25:53 crc kubenswrapper[4931]: E1129 00:25:53.694769 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64db2499-e411-42b9-8e99-92e748a9ad5d" containerName="swift-ring-rebalance" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.694787 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="64db2499-e411-42b9-8e99-92e748a9ad5d" containerName="swift-ring-rebalance" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.695026 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="64db2499-e411-42b9-8e99-92e748a9ad5d" containerName="swift-ring-rebalance" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.696901 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.715574 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cvwd7"] Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.808830 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6e1b0b5-fb10-49f4-8e81-541921765521-catalog-content\") pod \"community-operators-cvwd7\" (UID: \"e6e1b0b5-fb10-49f4-8e81-541921765521\") " pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.809090 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs9pq\" (UniqueName: \"kubernetes.io/projected/e6e1b0b5-fb10-49f4-8e81-541921765521-kube-api-access-cs9pq\") pod \"community-operators-cvwd7\" (UID: \"e6e1b0b5-fb10-49f4-8e81-541921765521\") " pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.809237 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6e1b0b5-fb10-49f4-8e81-541921765521-utilities\") pod \"community-operators-cvwd7\" (UID: \"e6e1b0b5-fb10-49f4-8e81-541921765521\") " pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.911510 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6e1b0b5-fb10-49f4-8e81-541921765521-catalog-content\") pod \"community-operators-cvwd7\" (UID: \"e6e1b0b5-fb10-49f4-8e81-541921765521\") " pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.911649 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs9pq\" (UniqueName: \"kubernetes.io/projected/e6e1b0b5-fb10-49f4-8e81-541921765521-kube-api-access-cs9pq\") pod \"community-operators-cvwd7\" (UID: \"e6e1b0b5-fb10-49f4-8e81-541921765521\") " pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.911779 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6e1b0b5-fb10-49f4-8e81-541921765521-utilities\") pod \"community-operators-cvwd7\" (UID: \"e6e1b0b5-fb10-49f4-8e81-541921765521\") " pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.912253 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6e1b0b5-fb10-49f4-8e81-541921765521-catalog-content\") pod \"community-operators-cvwd7\" (UID: \"e6e1b0b5-fb10-49f4-8e81-541921765521\") " pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.912262 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6e1b0b5-fb10-49f4-8e81-541921765521-utilities\") pod \"community-operators-cvwd7\" (UID: \"e6e1b0b5-fb10-49f4-8e81-541921765521\") " pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:25:53 crc kubenswrapper[4931]: I1129 00:25:53.938612 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs9pq\" (UniqueName: \"kubernetes.io/projected/e6e1b0b5-fb10-49f4-8e81-541921765521-kube-api-access-cs9pq\") pod \"community-operators-cvwd7\" (UID: \"e6e1b0b5-fb10-49f4-8e81-541921765521\") " pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:25:54 crc kubenswrapper[4931]: I1129 00:25:54.024152 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:25:54 crc kubenswrapper[4931]: I1129 00:25:54.589142 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cvwd7"] Nov 29 00:25:55 crc kubenswrapper[4931]: I1129 00:25:55.008371 4931 generic.go:334] "Generic (PLEG): container finished" podID="e6e1b0b5-fb10-49f4-8e81-541921765521" containerID="4da687569f11bd289c4ac23056dd880065fba935770b804628f1af365f936fd8" exitCode=0 Nov 29 00:25:55 crc kubenswrapper[4931]: I1129 00:25:55.008458 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvwd7" event={"ID":"e6e1b0b5-fb10-49f4-8e81-541921765521","Type":"ContainerDied","Data":"4da687569f11bd289c4ac23056dd880065fba935770b804628f1af365f936fd8"} Nov 29 00:25:55 crc kubenswrapper[4931]: I1129 00:25:55.008534 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvwd7" event={"ID":"e6e1b0b5-fb10-49f4-8e81-541921765521","Type":"ContainerStarted","Data":"472273f34f312357853752f6e65c46122a028c196385a4299afbce960c679b37"} Nov 29 00:25:57 crc kubenswrapper[4931]: I1129 00:25:57.029456 4931 generic.go:334] "Generic (PLEG): container finished" podID="e6e1b0b5-fb10-49f4-8e81-541921765521" containerID="0049eec51597a45715980c7083c1acb4b7b7f1cafe1354c1d0bd5a614b2cd846" exitCode=0 Nov 29 00:25:57 crc kubenswrapper[4931]: I1129 00:25:57.029597 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvwd7" event={"ID":"e6e1b0b5-fb10-49f4-8e81-541921765521","Type":"ContainerDied","Data":"0049eec51597a45715980c7083c1acb4b7b7f1cafe1354c1d0bd5a614b2cd846"} Nov 29 00:25:57 crc kubenswrapper[4931]: E1129 00:25:57.646777 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-storage-0" podUID="a720d879-8b09-452d-8371-9bf75f28cbde" Nov 29 00:25:58 crc kubenswrapper[4931]: I1129 00:25:58.041103 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 29 00:25:58 crc kubenswrapper[4931]: I1129 00:25:58.042078 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvwd7" event={"ID":"e6e1b0b5-fb10-49f4-8e81-541921765521","Type":"ContainerStarted","Data":"ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30"} Nov 29 00:25:58 crc kubenswrapper[4931]: I1129 00:25:58.066731 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cvwd7" podStartSLOduration=2.6586718830000002 podStartE2EDuration="5.066710722s" podCreationTimestamp="2025-11-29 00:25:53 +0000 UTC" firstStartedPulling="2025-11-29 00:25:55.009750262 +0000 UTC m=+1568.171643494" lastFinishedPulling="2025-11-29 00:25:57.417789101 +0000 UTC m=+1570.579682333" observedRunningTime="2025-11-29 00:25:58.06139206 +0000 UTC m=+1571.223285302" watchObservedRunningTime="2025-11-29 00:25:58.066710722 +0000 UTC m=+1571.228603964" Nov 29 00:25:58 crc kubenswrapper[4931]: I1129 00:25:58.092590 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:25:58 crc kubenswrapper[4931]: I1129 00:25:58.101834 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a720d879-8b09-452d-8371-9bf75f28cbde-etc-swift\") pod \"swift-storage-0\" (UID: \"a720d879-8b09-452d-8371-9bf75f28cbde\") " pod="openstack/swift-storage-0" Nov 29 00:25:58 crc kubenswrapper[4931]: I1129 00:25:58.343238 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 29 00:25:58 crc kubenswrapper[4931]: W1129 00:25:58.870450 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda720d879_8b09_452d_8371_9bf75f28cbde.slice/crio-d4c3c119f8a26e7dfca0a42cf206de7e20f2bd5e88c93c4aa0cdc0f902ed182d WatchSource:0}: Error finding container d4c3c119f8a26e7dfca0a42cf206de7e20f2bd5e88c93c4aa0cdc0f902ed182d: Status 404 returned error can't find the container with id d4c3c119f8a26e7dfca0a42cf206de7e20f2bd5e88c93c4aa0cdc0f902ed182d Nov 29 00:25:58 crc kubenswrapper[4931]: I1129 00:25:58.872392 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 29 00:25:59 crc kubenswrapper[4931]: I1129 00:25:59.052436 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"d4c3c119f8a26e7dfca0a42cf206de7e20f2bd5e88c93c4aa0cdc0f902ed182d"} Nov 29 00:26:01 crc kubenswrapper[4931]: I1129 00:26:01.086279 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"8c496f59bd0bd095dc3d05dfd2aab8b4a7f5dcdcb2800eac771c8a1fdb29a419"} Nov 29 00:26:01 crc kubenswrapper[4931]: I1129 00:26:01.087473 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"6e73a1da6b3616351d7bd3d85c96bc80a3d6f69b45c1ef850028f495b936c6d2"} Nov 29 00:26:02 crc kubenswrapper[4931]: I1129 00:26:02.103244 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"daa03ec34187e1fe3dd2a01e94a99f12a631e6548d38d5efa031ec657ead2af2"} Nov 29 00:26:03 crc kubenswrapper[4931]: I1129 00:26:03.115802 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"72a9188ca21dafe35bfdc611eb703884ee1444d7e763a7836e252acc01eef83f"} Nov 29 00:26:04 crc kubenswrapper[4931]: I1129 00:26:04.024599 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:26:04 crc kubenswrapper[4931]: I1129 00:26:04.024969 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:26:04 crc kubenswrapper[4931]: I1129 00:26:04.091209 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:26:04 crc kubenswrapper[4931]: I1129 00:26:04.175607 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:26:04 crc kubenswrapper[4931]: I1129 00:26:04.346946 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cvwd7"] Nov 29 00:26:05 crc kubenswrapper[4931]: I1129 00:26:05.141668 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"a051b07b429ace743c1bb2e47a252940c7f3b2ea2126be8330d58589e574f1c2"} Nov 29 00:26:06 crc kubenswrapper[4931]: I1129 00:26:06.158252 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"9090d3fdb06fb50d05e1e9ceeb25014e435d766f6423f2d4b74cd52bdacebbc7"} Nov 29 00:26:06 crc kubenswrapper[4931]: I1129 00:26:06.158683 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"7dd20ed99364035cb22fc888b1c99ed9d3a5505c56c1ba4d0eb34b806c30c476"} Nov 29 00:26:06 crc kubenswrapper[4931]: I1129 00:26:06.158699 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"290dd92300c8e5f1a07887ae983cef024244435c43c2d5db62e19674ab8ce49f"} Nov 29 00:26:06 crc kubenswrapper[4931]: I1129 00:26:06.158442 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cvwd7" podUID="e6e1b0b5-fb10-49f4-8e81-541921765521" containerName="registry-server" containerID="cri-o://ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30" gracePeriod=2 Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.151265 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.167318 4931 generic.go:334] "Generic (PLEG): container finished" podID="e6e1b0b5-fb10-49f4-8e81-541921765521" containerID="ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30" exitCode=0 Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.167364 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvwd7" event={"ID":"e6e1b0b5-fb10-49f4-8e81-541921765521","Type":"ContainerDied","Data":"ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30"} Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.167370 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cvwd7" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.167397 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvwd7" event={"ID":"e6e1b0b5-fb10-49f4-8e81-541921765521","Type":"ContainerDied","Data":"472273f34f312357853752f6e65c46122a028c196385a4299afbce960c679b37"} Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.167455 4931 scope.go:117] "RemoveContainer" containerID="ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.204437 4931 scope.go:117] "RemoveContainer" containerID="0049eec51597a45715980c7083c1acb4b7b7f1cafe1354c1d0bd5a614b2cd846" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.226393 4931 scope.go:117] "RemoveContainer" containerID="4da687569f11bd289c4ac23056dd880065fba935770b804628f1af365f936fd8" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.266158 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6e1b0b5-fb10-49f4-8e81-541921765521-catalog-content\") pod \"e6e1b0b5-fb10-49f4-8e81-541921765521\" (UID: \"e6e1b0b5-fb10-49f4-8e81-541921765521\") " Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.266234 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs9pq\" (UniqueName: \"kubernetes.io/projected/e6e1b0b5-fb10-49f4-8e81-541921765521-kube-api-access-cs9pq\") pod \"e6e1b0b5-fb10-49f4-8e81-541921765521\" (UID: \"e6e1b0b5-fb10-49f4-8e81-541921765521\") " Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.266288 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6e1b0b5-fb10-49f4-8e81-541921765521-utilities\") pod \"e6e1b0b5-fb10-49f4-8e81-541921765521\" (UID: \"e6e1b0b5-fb10-49f4-8e81-541921765521\") " Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.267515 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6e1b0b5-fb10-49f4-8e81-541921765521-utilities" (OuterVolumeSpecName: "utilities") pod "e6e1b0b5-fb10-49f4-8e81-541921765521" (UID: "e6e1b0b5-fb10-49f4-8e81-541921765521"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.274090 4931 scope.go:117] "RemoveContainer" containerID="ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.274166 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6e1b0b5-fb10-49f4-8e81-541921765521-kube-api-access-cs9pq" (OuterVolumeSpecName: "kube-api-access-cs9pq") pod "e6e1b0b5-fb10-49f4-8e81-541921765521" (UID: "e6e1b0b5-fb10-49f4-8e81-541921765521"). InnerVolumeSpecName "kube-api-access-cs9pq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:26:07 crc kubenswrapper[4931]: E1129 00:26:07.276706 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30\": container with ID starting with ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30 not found: ID does not exist" containerID="ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.276759 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30"} err="failed to get container status \"ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30\": rpc error: code = NotFound desc = could not find container \"ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30\": container with ID starting with ca56559b304125ec66f138f413abe33c7f3fa15bd623588d7807fa82150d8d30 not found: ID does not exist" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.276789 4931 scope.go:117] "RemoveContainer" containerID="0049eec51597a45715980c7083c1acb4b7b7f1cafe1354c1d0bd5a614b2cd846" Nov 29 00:26:07 crc kubenswrapper[4931]: E1129 00:26:07.277555 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0049eec51597a45715980c7083c1acb4b7b7f1cafe1354c1d0bd5a614b2cd846\": container with ID starting with 0049eec51597a45715980c7083c1acb4b7b7f1cafe1354c1d0bd5a614b2cd846 not found: ID does not exist" containerID="0049eec51597a45715980c7083c1acb4b7b7f1cafe1354c1d0bd5a614b2cd846" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.277588 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0049eec51597a45715980c7083c1acb4b7b7f1cafe1354c1d0bd5a614b2cd846"} err="failed to get container status \"0049eec51597a45715980c7083c1acb4b7b7f1cafe1354c1d0bd5a614b2cd846\": rpc error: code = NotFound desc = could not find container \"0049eec51597a45715980c7083c1acb4b7b7f1cafe1354c1d0bd5a614b2cd846\": container with ID starting with 0049eec51597a45715980c7083c1acb4b7b7f1cafe1354c1d0bd5a614b2cd846 not found: ID does not exist" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.277606 4931 scope.go:117] "RemoveContainer" containerID="4da687569f11bd289c4ac23056dd880065fba935770b804628f1af365f936fd8" Nov 29 00:26:07 crc kubenswrapper[4931]: E1129 00:26:07.277898 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4da687569f11bd289c4ac23056dd880065fba935770b804628f1af365f936fd8\": container with ID starting with 4da687569f11bd289c4ac23056dd880065fba935770b804628f1af365f936fd8 not found: ID does not exist" containerID="4da687569f11bd289c4ac23056dd880065fba935770b804628f1af365f936fd8" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.277928 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4da687569f11bd289c4ac23056dd880065fba935770b804628f1af365f936fd8"} err="failed to get container status \"4da687569f11bd289c4ac23056dd880065fba935770b804628f1af365f936fd8\": rpc error: code = NotFound desc = could not find container \"4da687569f11bd289c4ac23056dd880065fba935770b804628f1af365f936fd8\": container with ID starting with 4da687569f11bd289c4ac23056dd880065fba935770b804628f1af365f936fd8 not found: ID does not exist" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.326898 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6e1b0b5-fb10-49f4-8e81-541921765521-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6e1b0b5-fb10-49f4-8e81-541921765521" (UID: "e6e1b0b5-fb10-49f4-8e81-541921765521"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.368744 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6e1b0b5-fb10-49f4-8e81-541921765521-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.368775 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs9pq\" (UniqueName: \"kubernetes.io/projected/e6e1b0b5-fb10-49f4-8e81-541921765521-kube-api-access-cs9pq\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.368785 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6e1b0b5-fb10-49f4-8e81-541921765521-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.513432 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cvwd7"] Nov 29 00:26:07 crc kubenswrapper[4931]: I1129 00:26:07.524586 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cvwd7"] Nov 29 00:26:08 crc kubenswrapper[4931]: I1129 00:26:08.190212 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"1984156b423120a7603ae79fa13b2da9cd5c443e1892ab027def80c6e27c2330"} Nov 29 00:26:08 crc kubenswrapper[4931]: I1129 00:26:08.190265 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"852fd875775869d02f4d9157a7c5c96ba9f9bd290fba49cf2348ac3af64b2f6b"} Nov 29 00:26:08 crc kubenswrapper[4931]: I1129 00:26:08.190275 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"4e426f36a55c7a0fc9dfe16c934fb7a7fc434040a38b73a03bf2f0e37042c9ff"} Nov 29 00:26:08 crc kubenswrapper[4931]: I1129 00:26:08.190283 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"e5663470f7b52960319eb932327c920e240599f1e7bd43b2f387ecd0a83e3c87"} Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.239888 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6e1b0b5-fb10-49f4-8e81-541921765521" path="/var/lib/kubelet/pods/e6e1b0b5-fb10-49f4-8e81-541921765521/volumes" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.241231 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"257e5e5a7c1155fa55beb152fcc824afc44ff7ddf6c46c0937ee1f0f2bb0d415"} Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.241278 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"48d66bd9f32cfbdbb6d01e872bf816cfe203449483f411fa8707475677c95dcc"} Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.241294 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a720d879-8b09-452d-8371-9bf75f28cbde","Type":"ContainerStarted","Data":"34954f41d468ce48f1dd8f247d18798b3ae8b8e2e6e22ac43f54c47b18e68f0e"} Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.289536 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=499.194123165 podStartE2EDuration="8m27.28951102s" podCreationTimestamp="2025-11-29 00:17:42 +0000 UTC" firstStartedPulling="2025-11-29 00:25:58.872728224 +0000 UTC m=+1572.034621456" lastFinishedPulling="2025-11-29 00:26:06.968116079 +0000 UTC m=+1580.130009311" observedRunningTime="2025-11-29 00:26:09.279920526 +0000 UTC m=+1582.441813828" watchObservedRunningTime="2025-11-29 00:26:09.28951102 +0000 UTC m=+1582.451404262" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.588330 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5fcbbbc747-hm8g8"] Nov 29 00:26:09 crc kubenswrapper[4931]: E1129 00:26:09.588729 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6e1b0b5-fb10-49f4-8e81-541921765521" containerName="extract-content" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.588750 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6e1b0b5-fb10-49f4-8e81-541921765521" containerName="extract-content" Nov 29 00:26:09 crc kubenswrapper[4931]: E1129 00:26:09.588786 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6e1b0b5-fb10-49f4-8e81-541921765521" containerName="extract-utilities" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.588795 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6e1b0b5-fb10-49f4-8e81-541921765521" containerName="extract-utilities" Nov 29 00:26:09 crc kubenswrapper[4931]: E1129 00:26:09.588830 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6e1b0b5-fb10-49f4-8e81-541921765521" containerName="registry-server" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.588839 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6e1b0b5-fb10-49f4-8e81-541921765521" containerName="registry-server" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.589100 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6e1b0b5-fb10-49f4-8e81-541921765521" containerName="registry-server" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.590477 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.594055 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.609591 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fcbbbc747-hm8g8"] Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.613922 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-dns-svc\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.614181 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-config\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.614247 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-ovsdbserver-sb\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.614433 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-ovsdbserver-nb\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.614514 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-dns-swift-storage-0\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.614586 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6jr6\" (UniqueName: \"kubernetes.io/projected/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-kube-api-access-b6jr6\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.716136 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-ovsdbserver-nb\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.716196 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-dns-swift-storage-0\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.716230 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6jr6\" (UniqueName: \"kubernetes.io/projected/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-kube-api-access-b6jr6\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.716291 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-dns-svc\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.716378 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-config\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.716398 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-ovsdbserver-sb\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.717126 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-dns-swift-storage-0\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.717292 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-ovsdbserver-sb\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.717953 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-config\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.718089 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-ovsdbserver-nb\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.718384 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-dns-svc\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.734585 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6jr6\" (UniqueName: \"kubernetes.io/projected/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-kube-api-access-b6jr6\") pod \"dnsmasq-dns-5fcbbbc747-hm8g8\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:09 crc kubenswrapper[4931]: I1129 00:26:09.910352 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:10 crc kubenswrapper[4931]: I1129 00:26:10.430091 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fcbbbc747-hm8g8"] Nov 29 00:26:10 crc kubenswrapper[4931]: W1129 00:26:10.435487 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ab2b8a0_0372_4856_8c5b_55b187fb8c0b.slice/crio-b6267185cfa2a52f6a2a59ed47fe1bf20c6e43e3fed296090f1b6d93411946c5 WatchSource:0}: Error finding container b6267185cfa2a52f6a2a59ed47fe1bf20c6e43e3fed296090f1b6d93411946c5: Status 404 returned error can't find the container with id b6267185cfa2a52f6a2a59ed47fe1bf20c6e43e3fed296090f1b6d93411946c5 Nov 29 00:26:11 crc kubenswrapper[4931]: I1129 00:26:11.250541 4931 generic.go:334] "Generic (PLEG): container finished" podID="8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" containerID="a4537a7cc0010a23bdf364b17f5e57721b8e67f64825a4190d9cb057c478e662" exitCode=0 Nov 29 00:26:11 crc kubenswrapper[4931]: I1129 00:26:11.250638 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" event={"ID":"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b","Type":"ContainerDied","Data":"a4537a7cc0010a23bdf364b17f5e57721b8e67f64825a4190d9cb057c478e662"} Nov 29 00:26:11 crc kubenswrapper[4931]: I1129 00:26:11.250869 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" event={"ID":"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b","Type":"ContainerStarted","Data":"b6267185cfa2a52f6a2a59ed47fe1bf20c6e43e3fed296090f1b6d93411946c5"} Nov 29 00:26:12 crc kubenswrapper[4931]: I1129 00:26:12.268963 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" event={"ID":"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b","Type":"ContainerStarted","Data":"5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f"} Nov 29 00:26:12 crc kubenswrapper[4931]: I1129 00:26:12.269295 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:12 crc kubenswrapper[4931]: I1129 00:26:12.296897 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" podStartSLOduration=3.296870452 podStartE2EDuration="3.296870452s" podCreationTimestamp="2025-11-29 00:26:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:26:12.295459012 +0000 UTC m=+1585.457352294" watchObservedRunningTime="2025-11-29 00:26:12.296870452 +0000 UTC m=+1585.458763714" Nov 29 00:26:13 crc kubenswrapper[4931]: I1129 00:26:13.761600 4931 scope.go:117] "RemoveContainer" containerID="4d5dc2f4eab8c21d8b7a78f5f16f4a198c7245fd51df7958bff094f5455d0450" Nov 29 00:26:13 crc kubenswrapper[4931]: I1129 00:26:13.796179 4931 scope.go:117] "RemoveContainer" containerID="a8fda88fd35b023ddcfa5bdd32d75ccc4abad9b5fb0dffa414b56d66ba10dd4d" Nov 29 00:26:13 crc kubenswrapper[4931]: E1129 00:26:13.846388 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[etc-swift], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openstack/swift-proxy-764cb44577-7q67f" podUID="550e16f7-f101-4a91-8e2f-45af3430d6b9" Nov 29 00:26:14 crc kubenswrapper[4931]: I1129 00:26:14.291381 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:26:16 crc kubenswrapper[4931]: I1129 00:26:16.345420 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:26:16 crc kubenswrapper[4931]: I1129 00:26:16.373071 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/550e16f7-f101-4a91-8e2f-45af3430d6b9-etc-swift\") pod \"swift-proxy-764cb44577-7q67f\" (UID: \"550e16f7-f101-4a91-8e2f-45af3430d6b9\") " pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:26:16 crc kubenswrapper[4931]: I1129 00:26:16.393252 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:26:16 crc kubenswrapper[4931]: I1129 00:26:16.948430 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-764cb44577-7q67f"] Nov 29 00:26:17 crc kubenswrapper[4931]: I1129 00:26:17.328491 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-764cb44577-7q67f" event={"ID":"550e16f7-f101-4a91-8e2f-45af3430d6b9","Type":"ContainerStarted","Data":"1fbd4fa74fc7ace941ba6c987927ec3de7e39d886c5d458d5eb2efcfb3f2b72d"} Nov 29 00:26:17 crc kubenswrapper[4931]: I1129 00:26:17.328844 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-764cb44577-7q67f" event={"ID":"550e16f7-f101-4a91-8e2f-45af3430d6b9","Type":"ContainerStarted","Data":"98f4e63c698cbcf778dec73b3b3b6cf4015fe9d9331e8d4b96fc6a331e82e0d3"} Nov 29 00:26:17 crc kubenswrapper[4931]: I1129 00:26:17.328865 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:26:17 crc kubenswrapper[4931]: I1129 00:26:17.328880 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-764cb44577-7q67f" event={"ID":"550e16f7-f101-4a91-8e2f-45af3430d6b9","Type":"ContainerStarted","Data":"61cc7ca6bcd6a60946e3409dd38d198add7cb78f58d8a3afc616d3d01d59c080"} Nov 29 00:26:17 crc kubenswrapper[4931]: I1129 00:26:17.347915 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-764cb44577-7q67f" podStartSLOduration=374.347898936 podStartE2EDuration="6m14.347898936s" podCreationTimestamp="2025-11-29 00:20:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:26:17.34313695 +0000 UTC m=+1590.505030202" watchObservedRunningTime="2025-11-29 00:26:17.347898936 +0000 UTC m=+1590.509792168" Nov 29 00:26:18 crc kubenswrapper[4931]: I1129 00:26:18.337536 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:26:19 crc kubenswrapper[4931]: I1129 00:26:19.912040 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:19 crc kubenswrapper[4931]: I1129 00:26:19.992868 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5459cb87c-zlwt7"] Nov 29 00:26:19 crc kubenswrapper[4931]: I1129 00:26:19.993181 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" podUID="7cc674cd-2e25-45be-acad-b6031e34bd24" containerName="dnsmasq-dns" containerID="cri-o://b13f34346ffb8d9a8b28f94175c7e06620a285a936c46c094f87b08bf6286b72" gracePeriod=10 Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.355562 4931 generic.go:334] "Generic (PLEG): container finished" podID="7cc674cd-2e25-45be-acad-b6031e34bd24" containerID="b13f34346ffb8d9a8b28f94175c7e06620a285a936c46c094f87b08bf6286b72" exitCode=0 Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.356009 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" event={"ID":"7cc674cd-2e25-45be-acad-b6031e34bd24","Type":"ContainerDied","Data":"b13f34346ffb8d9a8b28f94175c7e06620a285a936c46c094f87b08bf6286b72"} Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.526745 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.730428 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x788c\" (UniqueName: \"kubernetes.io/projected/7cc674cd-2e25-45be-acad-b6031e34bd24-kube-api-access-x788c\") pod \"7cc674cd-2e25-45be-acad-b6031e34bd24\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.730496 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-ovsdbserver-nb\") pod \"7cc674cd-2e25-45be-acad-b6031e34bd24\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.730555 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-config\") pod \"7cc674cd-2e25-45be-acad-b6031e34bd24\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.730633 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-ovsdbserver-sb\") pod \"7cc674cd-2e25-45be-acad-b6031e34bd24\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.730688 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-dns-svc\") pod \"7cc674cd-2e25-45be-acad-b6031e34bd24\" (UID: \"7cc674cd-2e25-45be-acad-b6031e34bd24\") " Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.744193 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cc674cd-2e25-45be-acad-b6031e34bd24-kube-api-access-x788c" (OuterVolumeSpecName: "kube-api-access-x788c") pod "7cc674cd-2e25-45be-acad-b6031e34bd24" (UID: "7cc674cd-2e25-45be-acad-b6031e34bd24"). InnerVolumeSpecName "kube-api-access-x788c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.794075 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7cc674cd-2e25-45be-acad-b6031e34bd24" (UID: "7cc674cd-2e25-45be-acad-b6031e34bd24"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.801351 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7cc674cd-2e25-45be-acad-b6031e34bd24" (UID: "7cc674cd-2e25-45be-acad-b6031e34bd24"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.812924 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-config" (OuterVolumeSpecName: "config") pod "7cc674cd-2e25-45be-acad-b6031e34bd24" (UID: "7cc674cd-2e25-45be-acad-b6031e34bd24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.818608 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7cc674cd-2e25-45be-acad-b6031e34bd24" (UID: "7cc674cd-2e25-45be-acad-b6031e34bd24"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.833106 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.833153 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x788c\" (UniqueName: \"kubernetes.io/projected/7cc674cd-2e25-45be-acad-b6031e34bd24-kube-api-access-x788c\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.833167 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.833186 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:20.833198 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cc674cd-2e25-45be-acad-b6031e34bd24-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:21.367974 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" event={"ID":"7cc674cd-2e25-45be-acad-b6031e34bd24","Type":"ContainerDied","Data":"712e1de4433955f789576af7e20c641d1b577ee30ceba6af9aff8d6c385da297"} Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:21.368038 4931 scope.go:117] "RemoveContainer" containerID="b13f34346ffb8d9a8b28f94175c7e06620a285a936c46c094f87b08bf6286b72" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:21.368084 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5459cb87c-zlwt7" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:21.396677 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5459cb87c-zlwt7"] Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:21.406407 4931 scope.go:117] "RemoveContainer" containerID="2138484b45b0f7b2da55011d8546db5b8dc8893b5a2e275c10ca97b22e65745a" Nov 29 00:26:21 crc kubenswrapper[4931]: I1129 00:26:21.418394 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5459cb87c-zlwt7"] Nov 29 00:26:23 crc kubenswrapper[4931]: I1129 00:26:23.233802 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cc674cd-2e25-45be-acad-b6031e34bd24" path="/var/lib/kubelet/pods/7cc674cd-2e25-45be-acad-b6031e34bd24/volumes" Nov 29 00:26:26 crc kubenswrapper[4931]: I1129 00:26:26.401383 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:26:26 crc kubenswrapper[4931]: I1129 00:26:26.404469 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-764cb44577-7q67f" Nov 29 00:26:35 crc kubenswrapper[4931]: I1129 00:26:35.172099 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 00:26:36 crc kubenswrapper[4931]: I1129 00:26:36.109950 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 00:26:39 crc kubenswrapper[4931]: I1129 00:26:39.986589 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0d661705-7bb5-482f-b2a8-79b44f6351ac" containerName="rabbitmq" containerID="cri-o://e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77" gracePeriod=604796 Nov 29 00:26:40 crc kubenswrapper[4931]: I1129 00:26:40.071733 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="fbd78afc-2d6d-4118-92a2-e3499560359f" containerName="rabbitmq" containerID="cri-o://39fb76b6f4639dac5c87ee61fb096b14dda68bf6dce825287bbabae1121c3285" gracePeriod=604797 Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.572898 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66c8d98dc5-q25f9"] Nov 29 00:26:46 crc kubenswrapper[4931]: E1129 00:26:46.612152 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cc674cd-2e25-45be-acad-b6031e34bd24" containerName="init" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.612181 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cc674cd-2e25-45be-acad-b6031e34bd24" containerName="init" Nov 29 00:26:46 crc kubenswrapper[4931]: E1129 00:26:46.612201 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cc674cd-2e25-45be-acad-b6031e34bd24" containerName="dnsmasq-dns" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.612211 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cc674cd-2e25-45be-acad-b6031e34bd24" containerName="dnsmasq-dns" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.612468 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cc674cd-2e25-45be-acad-b6031e34bd24" containerName="dnsmasq-dns" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.613773 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66c8d98dc5-q25f9"] Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.613908 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.628272 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.664966 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tksjj\" (UniqueName: \"kubernetes.io/projected/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-kube-api-access-tksjj\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.665040 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-openstack-edpm-ipam\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.665060 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-dns-swift-storage-0\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.665089 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-dns-svc\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.665121 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-ovsdbserver-nb\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.665167 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-config\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.665210 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-ovsdbserver-sb\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.676958 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.680555 4931 generic.go:334] "Generic (PLEG): container finished" podID="0d661705-7bb5-482f-b2a8-79b44f6351ac" containerID="e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77" exitCode=0 Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.680628 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0d661705-7bb5-482f-b2a8-79b44f6351ac","Type":"ContainerDied","Data":"e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77"} Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.680656 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0d661705-7bb5-482f-b2a8-79b44f6351ac","Type":"ContainerDied","Data":"4e5115ada1c17ffa4b9282df255656354f160bb6c4761063ca2718a5f5294d12"} Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.680674 4931 scope.go:117] "RemoveContainer" containerID="e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.693876 4931 generic.go:334] "Generic (PLEG): container finished" podID="fbd78afc-2d6d-4118-92a2-e3499560359f" containerID="39fb76b6f4639dac5c87ee61fb096b14dda68bf6dce825287bbabae1121c3285" exitCode=0 Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.693913 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fbd78afc-2d6d-4118-92a2-e3499560359f","Type":"ContainerDied","Data":"39fb76b6f4639dac5c87ee61fb096b14dda68bf6dce825287bbabae1121c3285"} Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.693937 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fbd78afc-2d6d-4118-92a2-e3499560359f","Type":"ContainerDied","Data":"d17cdeaa13e0bb877cdb7326736d7f10f567d69ed0d7fcdba99c86f897e3bc3c"} Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.693950 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d17cdeaa13e0bb877cdb7326736d7f10f567d69ed0d7fcdba99c86f897e3bc3c" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.694051 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.754884 4931 scope.go:117] "RemoveContainer" containerID="c13dfad395ef5b782b6371d50ee00ee1d0b2fdc386adbf7dfeb5c206bbf2019d" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.768630 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"fbd78afc-2d6d-4118-92a2-e3499560359f\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.768699 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-tls\") pod \"fbd78afc-2d6d-4118-92a2-e3499560359f\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.768724 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-config-data\") pod \"0d661705-7bb5-482f-b2a8-79b44f6351ac\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.768751 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fbd78afc-2d6d-4118-92a2-e3499560359f-pod-info\") pod \"fbd78afc-2d6d-4118-92a2-e3499560359f\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.768782 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-confd\") pod \"fbd78afc-2d6d-4118-92a2-e3499560359f\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.768830 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-plugins-conf\") pod \"0d661705-7bb5-482f-b2a8-79b44f6351ac\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.768867 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-erlang-cookie\") pod \"0d661705-7bb5-482f-b2a8-79b44f6351ac\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.768936 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-config-data\") pod \"fbd78afc-2d6d-4118-92a2-e3499560359f\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.768985 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-erlang-cookie\") pod \"fbd78afc-2d6d-4118-92a2-e3499560359f\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769007 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-plugins\") pod \"0d661705-7bb5-482f-b2a8-79b44f6351ac\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769041 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-tls\") pod \"0d661705-7bb5-482f-b2a8-79b44f6351ac\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769064 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-plugins\") pod \"fbd78afc-2d6d-4118-92a2-e3499560359f\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769114 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-confd\") pod \"0d661705-7bb5-482f-b2a8-79b44f6351ac\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769157 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-server-conf\") pod \"fbd78afc-2d6d-4118-92a2-e3499560359f\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769177 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hhmg\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-kube-api-access-2hhmg\") pod \"0d661705-7bb5-482f-b2a8-79b44f6351ac\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769217 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0d661705-7bb5-482f-b2a8-79b44f6351ac-pod-info\") pod \"0d661705-7bb5-482f-b2a8-79b44f6351ac\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769239 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-plugins-conf\") pod \"fbd78afc-2d6d-4118-92a2-e3499560359f\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769259 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"0d661705-7bb5-482f-b2a8-79b44f6351ac\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769289 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-server-conf\") pod \"0d661705-7bb5-482f-b2a8-79b44f6351ac\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769316 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fbd78afc-2d6d-4118-92a2-e3499560359f-erlang-cookie-secret\") pod \"fbd78afc-2d6d-4118-92a2-e3499560359f\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769349 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0d661705-7bb5-482f-b2a8-79b44f6351ac-erlang-cookie-secret\") pod \"0d661705-7bb5-482f-b2a8-79b44f6351ac\" (UID: \"0d661705-7bb5-482f-b2a8-79b44f6351ac\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769387 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwjd9\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-kube-api-access-qwjd9\") pod \"fbd78afc-2d6d-4118-92a2-e3499560359f\" (UID: \"fbd78afc-2d6d-4118-92a2-e3499560359f\") " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769644 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tksjj\" (UniqueName: \"kubernetes.io/projected/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-kube-api-access-tksjj\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769704 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-openstack-edpm-ipam\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769728 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-dns-swift-storage-0\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769775 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-dns-svc\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769828 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-ovsdbserver-nb\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769882 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-config\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.769926 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-ovsdbserver-sb\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.770900 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-ovsdbserver-sb\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.772361 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "fbd78afc-2d6d-4118-92a2-e3499560359f" (UID: "fbd78afc-2d6d-4118-92a2-e3499560359f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.775674 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0d661705-7bb5-482f-b2a8-79b44f6351ac" (UID: "0d661705-7bb5-482f-b2a8-79b44f6351ac"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.776016 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "fbd78afc-2d6d-4118-92a2-e3499560359f" (UID: "fbd78afc-2d6d-4118-92a2-e3499560359f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.776314 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbd78afc-2d6d-4118-92a2-e3499560359f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "fbd78afc-2d6d-4118-92a2-e3499560359f" (UID: "fbd78afc-2d6d-4118-92a2-e3499560359f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.776556 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0d661705-7bb5-482f-b2a8-79b44f6351ac" (UID: "0d661705-7bb5-482f-b2a8-79b44f6351ac"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.776615 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "fbd78afc-2d6d-4118-92a2-e3499560359f" (UID: "fbd78afc-2d6d-4118-92a2-e3499560359f"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.779568 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0d661705-7bb5-482f-b2a8-79b44f6351ac-pod-info" (OuterVolumeSpecName: "pod-info") pod "0d661705-7bb5-482f-b2a8-79b44f6351ac" (UID: "0d661705-7bb5-482f-b2a8-79b44f6351ac"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.781594 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/fbd78afc-2d6d-4118-92a2-e3499560359f-pod-info" (OuterVolumeSpecName: "pod-info") pod "fbd78afc-2d6d-4118-92a2-e3499560359f" (UID: "fbd78afc-2d6d-4118-92a2-e3499560359f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.783977 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d661705-7bb5-482f-b2a8-79b44f6351ac-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0d661705-7bb5-482f-b2a8-79b44f6351ac" (UID: "0d661705-7bb5-482f-b2a8-79b44f6351ac"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.784142 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0d661705-7bb5-482f-b2a8-79b44f6351ac" (UID: "0d661705-7bb5-482f-b2a8-79b44f6351ac"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.784226 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-kube-api-access-2hhmg" (OuterVolumeSpecName: "kube-api-access-2hhmg") pod "0d661705-7bb5-482f-b2a8-79b44f6351ac" (UID: "0d661705-7bb5-482f-b2a8-79b44f6351ac"). InnerVolumeSpecName "kube-api-access-2hhmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.784876 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-dns-swift-storage-0\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.791222 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "fbd78afc-2d6d-4118-92a2-e3499560359f" (UID: "fbd78afc-2d6d-4118-92a2-e3499560359f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.791353 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-ovsdbserver-nb\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.791877 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-config\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.793599 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "fbd78afc-2d6d-4118-92a2-e3499560359f" (UID: "fbd78afc-2d6d-4118-92a2-e3499560359f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.794165 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0d661705-7bb5-482f-b2a8-79b44f6351ac" (UID: "0d661705-7bb5-482f-b2a8-79b44f6351ac"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.795997 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "0d661705-7bb5-482f-b2a8-79b44f6351ac" (UID: "0d661705-7bb5-482f-b2a8-79b44f6351ac"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.810516 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-openstack-edpm-ipam\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.812392 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-dns-svc\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.813222 4931 scope.go:117] "RemoveContainer" containerID="e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77" Nov 29 00:26:46 crc kubenswrapper[4931]: E1129 00:26:46.814051 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77\": container with ID starting with e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77 not found: ID does not exist" containerID="e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.814092 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77"} err="failed to get container status \"e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77\": rpc error: code = NotFound desc = could not find container \"e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77\": container with ID starting with e74ccaeeaa9ea1b0dc7dc455eb34786667ee61fd79f411ef4d61e9f2d7243a77 not found: ID does not exist" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.814118 4931 scope.go:117] "RemoveContainer" containerID="c13dfad395ef5b782b6371d50ee00ee1d0b2fdc386adbf7dfeb5c206bbf2019d" Nov 29 00:26:46 crc kubenswrapper[4931]: E1129 00:26:46.816964 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c13dfad395ef5b782b6371d50ee00ee1d0b2fdc386adbf7dfeb5c206bbf2019d\": container with ID starting with c13dfad395ef5b782b6371d50ee00ee1d0b2fdc386adbf7dfeb5c206bbf2019d not found: ID does not exist" containerID="c13dfad395ef5b782b6371d50ee00ee1d0b2fdc386adbf7dfeb5c206bbf2019d" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.817021 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c13dfad395ef5b782b6371d50ee00ee1d0b2fdc386adbf7dfeb5c206bbf2019d"} err="failed to get container status \"c13dfad395ef5b782b6371d50ee00ee1d0b2fdc386adbf7dfeb5c206bbf2019d\": rpc error: code = NotFound desc = could not find container \"c13dfad395ef5b782b6371d50ee00ee1d0b2fdc386adbf7dfeb5c206bbf2019d\": container with ID starting with c13dfad395ef5b782b6371d50ee00ee1d0b2fdc386adbf7dfeb5c206bbf2019d not found: ID does not exist" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.818913 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tksjj\" (UniqueName: \"kubernetes.io/projected/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-kube-api-access-tksjj\") pod \"dnsmasq-dns-66c8d98dc5-q25f9\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.824245 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-kube-api-access-qwjd9" (OuterVolumeSpecName: "kube-api-access-qwjd9") pod "fbd78afc-2d6d-4118-92a2-e3499560359f" (UID: "fbd78afc-2d6d-4118-92a2-e3499560359f"). InnerVolumeSpecName "kube-api-access-qwjd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.848130 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-config-data" (OuterVolumeSpecName: "config-data") pod "fbd78afc-2d6d-4118-92a2-e3499560359f" (UID: "fbd78afc-2d6d-4118-92a2-e3499560359f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871754 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hhmg\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-kube-api-access-2hhmg\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871786 4931 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0d661705-7bb5-482f-b2a8-79b44f6351ac-pod-info\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871796 4931 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871827 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871838 4931 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fbd78afc-2d6d-4118-92a2-e3499560359f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871848 4931 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0d661705-7bb5-482f-b2a8-79b44f6351ac-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871857 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwjd9\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-kube-api-access-qwjd9\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871871 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871880 4931 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871891 4931 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fbd78afc-2d6d-4118-92a2-e3499560359f-pod-info\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871899 4931 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871909 4931 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871919 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871929 4931 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871937 4931 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871946 4931 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.871954 4931 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.889720 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-config-data" (OuterVolumeSpecName: "config-data") pod "0d661705-7bb5-482f-b2a8-79b44f6351ac" (UID: "0d661705-7bb5-482f-b2a8-79b44f6351ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.915189 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.934695 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.944512 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-server-conf" (OuterVolumeSpecName: "server-conf") pod "fbd78afc-2d6d-4118-92a2-e3499560359f" (UID: "fbd78afc-2d6d-4118-92a2-e3499560359f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.949246 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-server-conf" (OuterVolumeSpecName: "server-conf") pod "0d661705-7bb5-482f-b2a8-79b44f6351ac" (UID: "0d661705-7bb5-482f-b2a8-79b44f6351ac"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.966571 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0d661705-7bb5-482f-b2a8-79b44f6351ac" (UID: "0d661705-7bb5-482f-b2a8-79b44f6351ac"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.973996 4931 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0d661705-7bb5-482f-b2a8-79b44f6351ac-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.974032 4931 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fbd78afc-2d6d-4118-92a2-e3499560359f-server-conf\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.974043 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.974122 4931 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-server-conf\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.974134 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.974144 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d661705-7bb5-482f-b2a8-79b44f6351ac-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:46 crc kubenswrapper[4931]: I1129 00:26:46.990984 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.004378 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "fbd78afc-2d6d-4118-92a2-e3499560359f" (UID: "fbd78afc-2d6d-4118-92a2-e3499560359f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.075251 4931 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fbd78afc-2d6d-4118-92a2-e3499560359f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.438299 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66c8d98dc5-q25f9"] Nov 29 00:26:47 crc kubenswrapper[4931]: W1129 00:26:47.443325 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb39e8948_d6ea_4cca_8b5b_bc60019f7e03.slice/crio-890155735f5eecc1e2cbdc4917184ff5c0b9a9de4fdb5ef07b0aa96972ea024a WatchSource:0}: Error finding container 890155735f5eecc1e2cbdc4917184ff5c0b9a9de4fdb5ef07b0aa96972ea024a: Status 404 returned error can't find the container with id 890155735f5eecc1e2cbdc4917184ff5c0b9a9de4fdb5ef07b0aa96972ea024a Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.703682 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.705921 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.706145 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" event={"ID":"b39e8948-d6ea-4cca-8b5b-bc60019f7e03","Type":"ContainerStarted","Data":"890155735f5eecc1e2cbdc4917184ff5c0b9a9de4fdb5ef07b0aa96972ea024a"} Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.733727 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.744966 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.757543 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.768622 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.777464 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 00:26:47 crc kubenswrapper[4931]: E1129 00:26:47.778439 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d661705-7bb5-482f-b2a8-79b44f6351ac" containerName="setup-container" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.778499 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d661705-7bb5-482f-b2a8-79b44f6351ac" containerName="setup-container" Nov 29 00:26:47 crc kubenswrapper[4931]: E1129 00:26:47.778528 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d661705-7bb5-482f-b2a8-79b44f6351ac" containerName="rabbitmq" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.778536 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d661705-7bb5-482f-b2a8-79b44f6351ac" containerName="rabbitmq" Nov 29 00:26:47 crc kubenswrapper[4931]: E1129 00:26:47.778551 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbd78afc-2d6d-4118-92a2-e3499560359f" containerName="setup-container" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.778559 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbd78afc-2d6d-4118-92a2-e3499560359f" containerName="setup-container" Nov 29 00:26:47 crc kubenswrapper[4931]: E1129 00:26:47.778594 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbd78afc-2d6d-4118-92a2-e3499560359f" containerName="rabbitmq" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.778601 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbd78afc-2d6d-4118-92a2-e3499560359f" containerName="rabbitmq" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.778799 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d661705-7bb5-482f-b2a8-79b44f6351ac" containerName="rabbitmq" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.778863 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbd78afc-2d6d-4118-92a2-e3499560359f" containerName="rabbitmq" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.780244 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.784369 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.784369 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-h4vzq" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.784467 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.784532 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.784572 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.784569 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.784682 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.792355 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.794158 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.797278 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.797510 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.797681 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.797931 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wz4dt" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.798130 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.798417 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.802299 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.804532 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.839265 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.893421 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8ccbdb2-db21-4fce-9111-b34a6800e238-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.893475 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8ccbdb2-db21-4fce-9111-b34a6800e238-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.893502 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.893524 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8ccbdb2-db21-4fce-9111-b34a6800e238-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.893554 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8ccbdb2-db21-4fce-9111-b34a6800e238-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.893597 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgrrg\" (UniqueName: \"kubernetes.io/projected/f8ccbdb2-db21-4fce-9111-b34a6800e238-kube-api-access-pgrrg\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.893623 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8ccbdb2-db21-4fce-9111-b34a6800e238-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.893659 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8ccbdb2-db21-4fce-9111-b34a6800e238-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.893680 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8ccbdb2-db21-4fce-9111-b34a6800e238-config-data\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.893718 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8ccbdb2-db21-4fce-9111-b34a6800e238-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.893744 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8ccbdb2-db21-4fce-9111-b34a6800e238-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995243 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgrrg\" (UniqueName: \"kubernetes.io/projected/f8ccbdb2-db21-4fce-9111-b34a6800e238-kube-api-access-pgrrg\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995308 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995336 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995373 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8ccbdb2-db21-4fce-9111-b34a6800e238-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995405 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995453 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njnqn\" (UniqueName: \"kubernetes.io/projected/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-kube-api-access-njnqn\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995483 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8ccbdb2-db21-4fce-9111-b34a6800e238-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995564 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8ccbdb2-db21-4fce-9111-b34a6800e238-config-data\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995649 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995696 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8ccbdb2-db21-4fce-9111-b34a6800e238-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995723 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8ccbdb2-db21-4fce-9111-b34a6800e238-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995758 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995797 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995850 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8ccbdb2-db21-4fce-9111-b34a6800e238-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995888 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8ccbdb2-db21-4fce-9111-b34a6800e238-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.995950 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.996020 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.996071 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8ccbdb2-db21-4fce-9111-b34a6800e238-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.996143 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.996176 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.996197 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8ccbdb2-db21-4fce-9111-b34a6800e238-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.996350 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.996366 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.996528 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f8ccbdb2-db21-4fce-9111-b34a6800e238-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.996998 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f8ccbdb2-db21-4fce-9111-b34a6800e238-config-data\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:47 crc kubenswrapper[4931]: I1129 00:26:47.997256 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f8ccbdb2-db21-4fce-9111-b34a6800e238-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.000045 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f8ccbdb2-db21-4fce-9111-b34a6800e238-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.000878 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f8ccbdb2-db21-4fce-9111-b34a6800e238-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.001887 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f8ccbdb2-db21-4fce-9111-b34a6800e238-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.004070 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f8ccbdb2-db21-4fce-9111-b34a6800e238-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.005146 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f8ccbdb2-db21-4fce-9111-b34a6800e238-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.005437 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f8ccbdb2-db21-4fce-9111-b34a6800e238-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.016751 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgrrg\" (UniqueName: \"kubernetes.io/projected/f8ccbdb2-db21-4fce-9111-b34a6800e238-kube-api-access-pgrrg\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.044980 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"f8ccbdb2-db21-4fce-9111-b34a6800e238\") " pod="openstack/rabbitmq-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.098634 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.098721 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.098766 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.098841 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njnqn\" (UniqueName: \"kubernetes.io/projected/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-kube-api-access-njnqn\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.098886 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.098956 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.099015 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.099102 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.099162 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.099204 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.099233 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.100411 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.101185 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.102366 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.102705 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.103440 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.103535 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.106078 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.106108 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.107529 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.108682 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.120538 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njnqn\" (UniqueName: \"kubernetes.io/projected/421b0e7c-7dcd-4e58-a2ee-445c18ec4b22-kube-api-access-njnqn\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.131327 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22\") " pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.161405 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.172896 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.674819 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.734116 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f8ccbdb2-db21-4fce-9111-b34a6800e238","Type":"ContainerStarted","Data":"e3d9ce9b28570c09f235ceedc03ab3d7a43f5929d61c36c35a6d3aec42b34255"} Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.735716 4931 generic.go:334] "Generic (PLEG): container finished" podID="b39e8948-d6ea-4cca-8b5b-bc60019f7e03" containerID="36dc249cbb98436f8441264051775a3dff3b764936dae8c504447dcc1860be4e" exitCode=0 Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.735763 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" event={"ID":"b39e8948-d6ea-4cca-8b5b-bc60019f7e03","Type":"ContainerDied","Data":"36dc249cbb98436f8441264051775a3dff3b764936dae8c504447dcc1860be4e"} Nov 29 00:26:48 crc kubenswrapper[4931]: W1129 00:26:48.908528 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod421b0e7c_7dcd_4e58_a2ee_445c18ec4b22.slice/crio-b5c7873e683dfc69fffc57f6d0a3e2a915136db4024a7da17d6e08a66c11e399 WatchSource:0}: Error finding container b5c7873e683dfc69fffc57f6d0a3e2a915136db4024a7da17d6e08a66c11e399: Status 404 returned error can't find the container with id b5c7873e683dfc69fffc57f6d0a3e2a915136db4024a7da17d6e08a66c11e399 Nov 29 00:26:48 crc kubenswrapper[4931]: I1129 00:26:48.909877 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 29 00:26:49 crc kubenswrapper[4931]: I1129 00:26:49.224862 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d661705-7bb5-482f-b2a8-79b44f6351ac" path="/var/lib/kubelet/pods/0d661705-7bb5-482f-b2a8-79b44f6351ac/volumes" Nov 29 00:26:49 crc kubenswrapper[4931]: I1129 00:26:49.226228 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbd78afc-2d6d-4118-92a2-e3499560359f" path="/var/lib/kubelet/pods/fbd78afc-2d6d-4118-92a2-e3499560359f/volumes" Nov 29 00:26:49 crc kubenswrapper[4931]: I1129 00:26:49.746043 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22","Type":"ContainerStarted","Data":"b5c7873e683dfc69fffc57f6d0a3e2a915136db4024a7da17d6e08a66c11e399"} Nov 29 00:26:49 crc kubenswrapper[4931]: I1129 00:26:49.747950 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" event={"ID":"b39e8948-d6ea-4cca-8b5b-bc60019f7e03","Type":"ContainerStarted","Data":"15cd65bfbcbc285064facf2ed3dfcf38c88d9787bfcecdf61b9d510cded88e35"} Nov 29 00:26:49 crc kubenswrapper[4931]: I1129 00:26:49.748096 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:49 crc kubenswrapper[4931]: I1129 00:26:49.779351 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" podStartSLOduration=3.779329105 podStartE2EDuration="3.779329105s" podCreationTimestamp="2025-11-29 00:26:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:26:49.76659341 +0000 UTC m=+1622.928486652" watchObservedRunningTime="2025-11-29 00:26:49.779329105 +0000 UTC m=+1622.941222327" Nov 29 00:26:50 crc kubenswrapper[4931]: I1129 00:26:50.759602 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f8ccbdb2-db21-4fce-9111-b34a6800e238","Type":"ContainerStarted","Data":"90405caa7a1a6e763a452e6e2aad630bf5577e65cd58e358775c05cb894f3347"} Nov 29 00:26:50 crc kubenswrapper[4931]: I1129 00:26:50.762529 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22","Type":"ContainerStarted","Data":"ccc46b12a25fb10acd7b1688d3a31c2ea9153a43ff8639ed6719c100aa7e89e0"} Nov 29 00:26:53 crc kubenswrapper[4931]: I1129 00:26:53.735588 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:26:53 crc kubenswrapper[4931]: I1129 00:26:53.735971 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:26:56 crc kubenswrapper[4931]: I1129 00:26:56.992985 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.074738 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fcbbbc747-hm8g8"] Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.077975 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" podUID="8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" containerName="dnsmasq-dns" containerID="cri-o://5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f" gracePeriod=10 Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.212765 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-578c4b6ff9-7b25t"] Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.215936 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.237013 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578c4b6ff9-7b25t"] Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.320229 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-ovsdbserver-nb\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.320285 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-config\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.320368 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-openstack-edpm-ipam\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.320392 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzgxv\" (UniqueName: \"kubernetes.io/projected/76e2de18-8995-446b-aaf0-427974e0334b-kube-api-access-xzgxv\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.320653 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-dns-swift-storage-0\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.320900 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-ovsdbserver-sb\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.320925 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-dns-svc\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.422786 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-config\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.422974 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-openstack-edpm-ipam\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.423005 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzgxv\" (UniqueName: \"kubernetes.io/projected/76e2de18-8995-446b-aaf0-427974e0334b-kube-api-access-xzgxv\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.423036 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-dns-swift-storage-0\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.423086 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-ovsdbserver-sb\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.423106 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-dns-svc\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.423138 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-ovsdbserver-nb\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.424026 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-openstack-edpm-ipam\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.424061 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-config\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.424230 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-dns-svc\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.424284 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-dns-swift-storage-0\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.424384 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-ovsdbserver-nb\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.424431 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76e2de18-8995-446b-aaf0-427974e0334b-ovsdbserver-sb\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.441681 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzgxv\" (UniqueName: \"kubernetes.io/projected/76e2de18-8995-446b-aaf0-427974e0334b-kube-api-access-xzgxv\") pod \"dnsmasq-dns-578c4b6ff9-7b25t\" (UID: \"76e2de18-8995-446b-aaf0-427974e0334b\") " pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.538311 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.557569 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.627742 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-ovsdbserver-nb\") pod \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.628112 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-dns-swift-storage-0\") pod \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.628141 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-config\") pod \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.628203 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-dns-svc\") pod \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.628230 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-ovsdbserver-sb\") pod \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.628362 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6jr6\" (UniqueName: \"kubernetes.io/projected/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-kube-api-access-b6jr6\") pod \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\" (UID: \"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b\") " Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.632151 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-kube-api-access-b6jr6" (OuterVolumeSpecName: "kube-api-access-b6jr6") pod "8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" (UID: "8ab2b8a0-0372-4856-8c5b-55b187fb8c0b"). InnerVolumeSpecName "kube-api-access-b6jr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.691912 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" (UID: "8ab2b8a0-0372-4856-8c5b-55b187fb8c0b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.694244 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" (UID: "8ab2b8a0-0372-4856-8c5b-55b187fb8c0b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.705264 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" (UID: "8ab2b8a0-0372-4856-8c5b-55b187fb8c0b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.720473 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" (UID: "8ab2b8a0-0372-4856-8c5b-55b187fb8c0b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.723289 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-config" (OuterVolumeSpecName: "config") pod "8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" (UID: "8ab2b8a0-0372-4856-8c5b-55b187fb8c0b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.731243 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.731278 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.731290 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.731305 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6jr6\" (UniqueName: \"kubernetes.io/projected/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-kube-api-access-b6jr6\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.731317 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.731325 4931 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.814041 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578c4b6ff9-7b25t"] Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.860295 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" event={"ID":"76e2de18-8995-446b-aaf0-427974e0334b","Type":"ContainerStarted","Data":"50fd4a296454d6b462595df053aba9b09d9ced7449e45919af423bcec020ba38"} Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.862251 4931 generic.go:334] "Generic (PLEG): container finished" podID="8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" containerID="5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f" exitCode=0 Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.862310 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.862306 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" event={"ID":"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b","Type":"ContainerDied","Data":"5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f"} Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.862529 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fcbbbc747-hm8g8" event={"ID":"8ab2b8a0-0372-4856-8c5b-55b187fb8c0b","Type":"ContainerDied","Data":"b6267185cfa2a52f6a2a59ed47fe1bf20c6e43e3fed296090f1b6d93411946c5"} Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.862551 4931 scope.go:117] "RemoveContainer" containerID="5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.889088 4931 scope.go:117] "RemoveContainer" containerID="a4537a7cc0010a23bdf364b17f5e57721b8e67f64825a4190d9cb057c478e662" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.900675 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fcbbbc747-hm8g8"] Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.916624 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5fcbbbc747-hm8g8"] Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.921885 4931 scope.go:117] "RemoveContainer" containerID="5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f" Nov 29 00:26:57 crc kubenswrapper[4931]: E1129 00:26:57.922307 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f\": container with ID starting with 5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f not found: ID does not exist" containerID="5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.922353 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f"} err="failed to get container status \"5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f\": rpc error: code = NotFound desc = could not find container \"5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f\": container with ID starting with 5020c1a7560b94071218a6cdeedc7bb4630466767340c9178f3a8c387306874f not found: ID does not exist" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.922385 4931 scope.go:117] "RemoveContainer" containerID="a4537a7cc0010a23bdf364b17f5e57721b8e67f64825a4190d9cb057c478e662" Nov 29 00:26:57 crc kubenswrapper[4931]: E1129 00:26:57.922863 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4537a7cc0010a23bdf364b17f5e57721b8e67f64825a4190d9cb057c478e662\": container with ID starting with a4537a7cc0010a23bdf364b17f5e57721b8e67f64825a4190d9cb057c478e662 not found: ID does not exist" containerID="a4537a7cc0010a23bdf364b17f5e57721b8e67f64825a4190d9cb057c478e662" Nov 29 00:26:57 crc kubenswrapper[4931]: I1129 00:26:57.922950 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4537a7cc0010a23bdf364b17f5e57721b8e67f64825a4190d9cb057c478e662"} err="failed to get container status \"a4537a7cc0010a23bdf364b17f5e57721b8e67f64825a4190d9cb057c478e662\": rpc error: code = NotFound desc = could not find container \"a4537a7cc0010a23bdf364b17f5e57721b8e67f64825a4190d9cb057c478e662\": container with ID starting with a4537a7cc0010a23bdf364b17f5e57721b8e67f64825a4190d9cb057c478e662 not found: ID does not exist" Nov 29 00:26:58 crc kubenswrapper[4931]: I1129 00:26:58.875134 4931 generic.go:334] "Generic (PLEG): container finished" podID="76e2de18-8995-446b-aaf0-427974e0334b" containerID="c2c1d442b7a124fab9a9b3e54900ddaeeecab1f2b294a93667ee261b56e0ec5b" exitCode=0 Nov 29 00:26:58 crc kubenswrapper[4931]: I1129 00:26:58.875231 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" event={"ID":"76e2de18-8995-446b-aaf0-427974e0334b","Type":"ContainerDied","Data":"c2c1d442b7a124fab9a9b3e54900ddaeeecab1f2b294a93667ee261b56e0ec5b"} Nov 29 00:26:59 crc kubenswrapper[4931]: I1129 00:26:59.220885 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" path="/var/lib/kubelet/pods/8ab2b8a0-0372-4856-8c5b-55b187fb8c0b/volumes" Nov 29 00:26:59 crc kubenswrapper[4931]: I1129 00:26:59.895344 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" event={"ID":"76e2de18-8995-446b-aaf0-427974e0334b","Type":"ContainerStarted","Data":"06770e8046ded89bcead513376d1c485b1e6db305fddf5b515e699e67f61197b"} Nov 29 00:26:59 crc kubenswrapper[4931]: I1129 00:26:59.895639 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:26:59 crc kubenswrapper[4931]: I1129 00:26:59.940288 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" podStartSLOduration=2.940258976 podStartE2EDuration="2.940258976s" podCreationTimestamp="2025-11-29 00:26:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:26:59.931263875 +0000 UTC m=+1633.093157177" watchObservedRunningTime="2025-11-29 00:26:59.940258976 +0000 UTC m=+1633.102152238" Nov 29 00:27:07 crc kubenswrapper[4931]: I1129 00:27:07.559113 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-578c4b6ff9-7b25t" Nov 29 00:27:07 crc kubenswrapper[4931]: I1129 00:27:07.673357 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66c8d98dc5-q25f9"] Nov 29 00:27:07 crc kubenswrapper[4931]: I1129 00:27:07.673679 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" podUID="b39e8948-d6ea-4cca-8b5b-bc60019f7e03" containerName="dnsmasq-dns" containerID="cri-o://15cd65bfbcbc285064facf2ed3dfcf38c88d9787bfcecdf61b9d510cded88e35" gracePeriod=10 Nov 29 00:27:07 crc kubenswrapper[4931]: I1129 00:27:07.974877 4931 generic.go:334] "Generic (PLEG): container finished" podID="b39e8948-d6ea-4cca-8b5b-bc60019f7e03" containerID="15cd65bfbcbc285064facf2ed3dfcf38c88d9787bfcecdf61b9d510cded88e35" exitCode=0 Nov 29 00:27:07 crc kubenswrapper[4931]: I1129 00:27:07.975018 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" event={"ID":"b39e8948-d6ea-4cca-8b5b-bc60019f7e03","Type":"ContainerDied","Data":"15cd65bfbcbc285064facf2ed3dfcf38c88d9787bfcecdf61b9d510cded88e35"} Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.312035 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.454602 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-ovsdbserver-nb\") pod \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.454658 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tksjj\" (UniqueName: \"kubernetes.io/projected/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-kube-api-access-tksjj\") pod \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.454736 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-dns-svc\") pod \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.454784 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-config\") pod \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.454861 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-dns-swift-storage-0\") pod \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.454895 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-openstack-edpm-ipam\") pod \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.454925 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-ovsdbserver-sb\") pod \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\" (UID: \"b39e8948-d6ea-4cca-8b5b-bc60019f7e03\") " Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.461186 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-kube-api-access-tksjj" (OuterVolumeSpecName: "kube-api-access-tksjj") pod "b39e8948-d6ea-4cca-8b5b-bc60019f7e03" (UID: "b39e8948-d6ea-4cca-8b5b-bc60019f7e03"). InnerVolumeSpecName "kube-api-access-tksjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.501092 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "b39e8948-d6ea-4cca-8b5b-bc60019f7e03" (UID: "b39e8948-d6ea-4cca-8b5b-bc60019f7e03"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.507241 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b39e8948-d6ea-4cca-8b5b-bc60019f7e03" (UID: "b39e8948-d6ea-4cca-8b5b-bc60019f7e03"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.512842 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b39e8948-d6ea-4cca-8b5b-bc60019f7e03" (UID: "b39e8948-d6ea-4cca-8b5b-bc60019f7e03"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.515017 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b39e8948-d6ea-4cca-8b5b-bc60019f7e03" (UID: "b39e8948-d6ea-4cca-8b5b-bc60019f7e03"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.516240 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-config" (OuterVolumeSpecName: "config") pod "b39e8948-d6ea-4cca-8b5b-bc60019f7e03" (UID: "b39e8948-d6ea-4cca-8b5b-bc60019f7e03"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.526371 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b39e8948-d6ea-4cca-8b5b-bc60019f7e03" (UID: "b39e8948-d6ea-4cca-8b5b-bc60019f7e03"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.557362 4931 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.557739 4931 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.557753 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.557765 4931 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.557775 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tksjj\" (UniqueName: \"kubernetes.io/projected/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-kube-api-access-tksjj\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.557791 4931 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.557801 4931 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b39e8948-d6ea-4cca-8b5b-bc60019f7e03-config\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.985967 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" event={"ID":"b39e8948-d6ea-4cca-8b5b-bc60019f7e03","Type":"ContainerDied","Data":"890155735f5eecc1e2cbdc4917184ff5c0b9a9de4fdb5ef07b0aa96972ea024a"} Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.986013 4931 scope.go:117] "RemoveContainer" containerID="15cd65bfbcbc285064facf2ed3dfcf38c88d9787bfcecdf61b9d510cded88e35" Nov 29 00:27:08 crc kubenswrapper[4931]: I1129 00:27:08.986148 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66c8d98dc5-q25f9" Nov 29 00:27:09 crc kubenswrapper[4931]: I1129 00:27:09.018726 4931 scope.go:117] "RemoveContainer" containerID="36dc249cbb98436f8441264051775a3dff3b764936dae8c504447dcc1860be4e" Nov 29 00:27:09 crc kubenswrapper[4931]: I1129 00:27:09.030511 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66c8d98dc5-q25f9"] Nov 29 00:27:09 crc kubenswrapper[4931]: I1129 00:27:09.041462 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66c8d98dc5-q25f9"] Nov 29 00:27:09 crc kubenswrapper[4931]: I1129 00:27:09.225423 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b39e8948-d6ea-4cca-8b5b-bc60019f7e03" path="/var/lib/kubelet/pods/b39e8948-d6ea-4cca-8b5b-bc60019f7e03/volumes" Nov 29 00:27:13 crc kubenswrapper[4931]: I1129 00:27:13.898995 4931 scope.go:117] "RemoveContainer" containerID="39fb76b6f4639dac5c87ee61fb096b14dda68bf6dce825287bbabae1121c3285" Nov 29 00:27:13 crc kubenswrapper[4931]: I1129 00:27:13.927267 4931 scope.go:117] "RemoveContainer" containerID="7fe4cd6379e10e7b3ca34aac9701c07da875d4602076ce55cb8db09358f686be" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.749874 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9"] Nov 29 00:27:20 crc kubenswrapper[4931]: E1129 00:27:20.750656 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" containerName="dnsmasq-dns" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.750669 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" containerName="dnsmasq-dns" Nov 29 00:27:20 crc kubenswrapper[4931]: E1129 00:27:20.750680 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b39e8948-d6ea-4cca-8b5b-bc60019f7e03" containerName="dnsmasq-dns" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.750685 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b39e8948-d6ea-4cca-8b5b-bc60019f7e03" containerName="dnsmasq-dns" Nov 29 00:27:20 crc kubenswrapper[4931]: E1129 00:27:20.750701 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b39e8948-d6ea-4cca-8b5b-bc60019f7e03" containerName="init" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.750709 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="b39e8948-d6ea-4cca-8b5b-bc60019f7e03" containerName="init" Nov 29 00:27:20 crc kubenswrapper[4931]: E1129 00:27:20.750717 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" containerName="init" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.750723 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" containerName="init" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.750900 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ab2b8a0-0372-4856-8c5b-55b187fb8c0b" containerName="dnsmasq-dns" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.750926 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="b39e8948-d6ea-4cca-8b5b-bc60019f7e03" containerName="dnsmasq-dns" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.751476 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.753364 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.753689 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.754051 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.754120 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.760962 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9"] Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.824794 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.824939 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.825531 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.825662 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqz9n\" (UniqueName: \"kubernetes.io/projected/8449a5b7-2fa6-4966-9d34-0243e3536f2b-kube-api-access-hqz9n\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.927705 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.927778 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqz9n\" (UniqueName: \"kubernetes.io/projected/8449a5b7-2fa6-4966-9d34-0243e3536f2b-kube-api-access-hqz9n\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.927836 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.927876 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.935208 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.939132 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.939593 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:20 crc kubenswrapper[4931]: I1129 00:27:20.949716 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqz9n\" (UniqueName: \"kubernetes.io/projected/8449a5b7-2fa6-4966-9d34-0243e3536f2b-kube-api-access-hqz9n\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:21 crc kubenswrapper[4931]: I1129 00:27:21.080013 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:21 crc kubenswrapper[4931]: I1129 00:27:21.643872 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9"] Nov 29 00:27:21 crc kubenswrapper[4931]: I1129 00:27:21.645964 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 00:27:22 crc kubenswrapper[4931]: I1129 00:27:22.131873 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" event={"ID":"8449a5b7-2fa6-4966-9d34-0243e3536f2b","Type":"ContainerStarted","Data":"4b1b523496f2429f0d1d2c9ea41def296cc86137f43fb625ca0452dd7155839e"} Nov 29 00:27:23 crc kubenswrapper[4931]: I1129 00:27:23.148442 4931 generic.go:334] "Generic (PLEG): container finished" podID="f8ccbdb2-db21-4fce-9111-b34a6800e238" containerID="90405caa7a1a6e763a452e6e2aad630bf5577e65cd58e358775c05cb894f3347" exitCode=0 Nov 29 00:27:23 crc kubenswrapper[4931]: I1129 00:27:23.148573 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f8ccbdb2-db21-4fce-9111-b34a6800e238","Type":"ContainerDied","Data":"90405caa7a1a6e763a452e6e2aad630bf5577e65cd58e358775c05cb894f3347"} Nov 29 00:27:23 crc kubenswrapper[4931]: I1129 00:27:23.734903 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:27:23 crc kubenswrapper[4931]: I1129 00:27:23.735176 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:27:24 crc kubenswrapper[4931]: I1129 00:27:24.159269 4931 generic.go:334] "Generic (PLEG): container finished" podID="421b0e7c-7dcd-4e58-a2ee-445c18ec4b22" containerID="ccc46b12a25fb10acd7b1688d3a31c2ea9153a43ff8639ed6719c100aa7e89e0" exitCode=0 Nov 29 00:27:24 crc kubenswrapper[4931]: I1129 00:27:24.159363 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22","Type":"ContainerDied","Data":"ccc46b12a25fb10acd7b1688d3a31c2ea9153a43ff8639ed6719c100aa7e89e0"} Nov 29 00:27:24 crc kubenswrapper[4931]: I1129 00:27:24.166913 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f8ccbdb2-db21-4fce-9111-b34a6800e238","Type":"ContainerStarted","Data":"8729b07df6dfd9b3ab7221af4d1eecf1ea538c50e46c34b7cf91fd736002f67b"} Nov 29 00:27:24 crc kubenswrapper[4931]: I1129 00:27:24.167902 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 29 00:27:24 crc kubenswrapper[4931]: I1129 00:27:24.262597 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.262569386 podStartE2EDuration="37.262569386s" podCreationTimestamp="2025-11-29 00:26:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:27:24.225473732 +0000 UTC m=+1657.387366964" watchObservedRunningTime="2025-11-29 00:27:24.262569386 +0000 UTC m=+1657.424462628" Nov 29 00:27:25 crc kubenswrapper[4931]: I1129 00:27:25.178047 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"421b0e7c-7dcd-4e58-a2ee-445c18ec4b22","Type":"ContainerStarted","Data":"e836c00d4da39dae678c14b319a18762790c9d1f302f5a9a0461377287c87f32"} Nov 29 00:27:25 crc kubenswrapper[4931]: I1129 00:27:25.179041 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:27:25 crc kubenswrapper[4931]: I1129 00:27:25.210550 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.210530722 podStartE2EDuration="38.210530722s" podCreationTimestamp="2025-11-29 00:26:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 00:27:25.204430762 +0000 UTC m=+1658.366324014" watchObservedRunningTime="2025-11-29 00:27:25.210530722 +0000 UTC m=+1658.372423954" Nov 29 00:27:31 crc kubenswrapper[4931]: I1129 00:27:31.239960 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" event={"ID":"8449a5b7-2fa6-4966-9d34-0243e3536f2b","Type":"ContainerStarted","Data":"58538404907b07fa7e01561194339873c890987ece090e782a6d8f164a6958ce"} Nov 29 00:27:31 crc kubenswrapper[4931]: I1129 00:27:31.276196 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" podStartSLOduration=2.371804688 podStartE2EDuration="11.276174771s" podCreationTimestamp="2025-11-29 00:27:20 +0000 UTC" firstStartedPulling="2025-11-29 00:27:21.645387505 +0000 UTC m=+1654.807280747" lastFinishedPulling="2025-11-29 00:27:30.549757598 +0000 UTC m=+1663.711650830" observedRunningTime="2025-11-29 00:27:31.268633381 +0000 UTC m=+1664.430526663" watchObservedRunningTime="2025-11-29 00:27:31.276174771 +0000 UTC m=+1664.438068023" Nov 29 00:27:38 crc kubenswrapper[4931]: I1129 00:27:38.165087 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 29 00:27:38 crc kubenswrapper[4931]: I1129 00:27:38.180059 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 29 00:27:42 crc kubenswrapper[4931]: I1129 00:27:42.383479 4931 generic.go:334] "Generic (PLEG): container finished" podID="8449a5b7-2fa6-4966-9d34-0243e3536f2b" containerID="58538404907b07fa7e01561194339873c890987ece090e782a6d8f164a6958ce" exitCode=0 Nov 29 00:27:42 crc kubenswrapper[4931]: I1129 00:27:42.383617 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" event={"ID":"8449a5b7-2fa6-4966-9d34-0243e3536f2b","Type":"ContainerDied","Data":"58538404907b07fa7e01561194339873c890987ece090e782a6d8f164a6958ce"} Nov 29 00:27:43 crc kubenswrapper[4931]: I1129 00:27:43.815401 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:43 crc kubenswrapper[4931]: I1129 00:27:43.935177 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqz9n\" (UniqueName: \"kubernetes.io/projected/8449a5b7-2fa6-4966-9d34-0243e3536f2b-kube-api-access-hqz9n\") pod \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " Nov 29 00:27:43 crc kubenswrapper[4931]: I1129 00:27:43.935377 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-repo-setup-combined-ca-bundle\") pod \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " Nov 29 00:27:43 crc kubenswrapper[4931]: I1129 00:27:43.935498 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-inventory\") pod \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " Nov 29 00:27:43 crc kubenswrapper[4931]: I1129 00:27:43.935572 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-ssh-key\") pod \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\" (UID: \"8449a5b7-2fa6-4966-9d34-0243e3536f2b\") " Nov 29 00:27:43 crc kubenswrapper[4931]: I1129 00:27:43.940880 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "8449a5b7-2fa6-4966-9d34-0243e3536f2b" (UID: "8449a5b7-2fa6-4966-9d34-0243e3536f2b"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:27:43 crc kubenswrapper[4931]: I1129 00:27:43.948373 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8449a5b7-2fa6-4966-9d34-0243e3536f2b-kube-api-access-hqz9n" (OuterVolumeSpecName: "kube-api-access-hqz9n") pod "8449a5b7-2fa6-4966-9d34-0243e3536f2b" (UID: "8449a5b7-2fa6-4966-9d34-0243e3536f2b"). InnerVolumeSpecName "kube-api-access-hqz9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:27:43 crc kubenswrapper[4931]: I1129 00:27:43.975758 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8449a5b7-2fa6-4966-9d34-0243e3536f2b" (UID: "8449a5b7-2fa6-4966-9d34-0243e3536f2b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:27:43 crc kubenswrapper[4931]: I1129 00:27:43.976093 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-inventory" (OuterVolumeSpecName: "inventory") pod "8449a5b7-2fa6-4966-9d34-0243e3536f2b" (UID: "8449a5b7-2fa6-4966-9d34-0243e3536f2b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.038157 4931 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.038211 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.038225 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8449a5b7-2fa6-4966-9d34-0243e3536f2b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.038234 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqz9n\" (UniqueName: \"kubernetes.io/projected/8449a5b7-2fa6-4966-9d34-0243e3536f2b-kube-api-access-hqz9n\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.417312 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" event={"ID":"8449a5b7-2fa6-4966-9d34-0243e3536f2b","Type":"ContainerDied","Data":"4b1b523496f2429f0d1d2c9ea41def296cc86137f43fb625ca0452dd7155839e"} Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.417460 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b1b523496f2429f0d1d2c9ea41def296cc86137f43fb625ca0452dd7155839e" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.417358 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.504327 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv"] Nov 29 00:27:44 crc kubenswrapper[4931]: E1129 00:27:44.504893 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8449a5b7-2fa6-4966-9d34-0243e3536f2b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.504922 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="8449a5b7-2fa6-4966-9d34-0243e3536f2b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.505149 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="8449a5b7-2fa6-4966-9d34-0243e3536f2b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.506141 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.508980 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.508981 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.509071 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.509762 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.542916 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv"] Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.650510 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2921aa97-795f-434a-ac6e-009e6555e864-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6rmv\" (UID: \"2921aa97-795f-434a-ac6e-009e6555e864\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.650931 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2921aa97-795f-434a-ac6e-009e6555e864-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6rmv\" (UID: \"2921aa97-795f-434a-ac6e-009e6555e864\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.651130 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22l6w\" (UniqueName: \"kubernetes.io/projected/2921aa97-795f-434a-ac6e-009e6555e864-kube-api-access-22l6w\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6rmv\" (UID: \"2921aa97-795f-434a-ac6e-009e6555e864\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.752878 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2921aa97-795f-434a-ac6e-009e6555e864-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6rmv\" (UID: \"2921aa97-795f-434a-ac6e-009e6555e864\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.753048 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2921aa97-795f-434a-ac6e-009e6555e864-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6rmv\" (UID: \"2921aa97-795f-434a-ac6e-009e6555e864\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.753128 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22l6w\" (UniqueName: \"kubernetes.io/projected/2921aa97-795f-434a-ac6e-009e6555e864-kube-api-access-22l6w\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6rmv\" (UID: \"2921aa97-795f-434a-ac6e-009e6555e864\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.764665 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2921aa97-795f-434a-ac6e-009e6555e864-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6rmv\" (UID: \"2921aa97-795f-434a-ac6e-009e6555e864\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.766349 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2921aa97-795f-434a-ac6e-009e6555e864-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6rmv\" (UID: \"2921aa97-795f-434a-ac6e-009e6555e864\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.785680 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22l6w\" (UniqueName: \"kubernetes.io/projected/2921aa97-795f-434a-ac6e-009e6555e864-kube-api-access-22l6w\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-n6rmv\" (UID: \"2921aa97-795f-434a-ac6e-009e6555e864\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:44 crc kubenswrapper[4931]: I1129 00:27:44.832871 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:45 crc kubenswrapper[4931]: I1129 00:27:45.407184 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv"] Nov 29 00:27:45 crc kubenswrapper[4931]: W1129 00:27:45.407347 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2921aa97_795f_434a_ac6e_009e6555e864.slice/crio-13a5ff8d3071f36dfc4f102fce50ca1b3f356983c1771b35ef49c39b1275003d WatchSource:0}: Error finding container 13a5ff8d3071f36dfc4f102fce50ca1b3f356983c1771b35ef49c39b1275003d: Status 404 returned error can't find the container with id 13a5ff8d3071f36dfc4f102fce50ca1b3f356983c1771b35ef49c39b1275003d Nov 29 00:27:45 crc kubenswrapper[4931]: I1129 00:27:45.427795 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" event={"ID":"2921aa97-795f-434a-ac6e-009e6555e864","Type":"ContainerStarted","Data":"13a5ff8d3071f36dfc4f102fce50ca1b3f356983c1771b35ef49c39b1275003d"} Nov 29 00:27:46 crc kubenswrapper[4931]: I1129 00:27:46.441003 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" event={"ID":"2921aa97-795f-434a-ac6e-009e6555e864","Type":"ContainerStarted","Data":"a771a225b7abc6155350e7c6ccf53c7cf1ca505d8a17f1acd5249bd91879c1c5"} Nov 29 00:27:46 crc kubenswrapper[4931]: I1129 00:27:46.476977 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" podStartSLOduration=2.026492513 podStartE2EDuration="2.476959107s" podCreationTimestamp="2025-11-29 00:27:44 +0000 UTC" firstStartedPulling="2025-11-29 00:27:45.409699645 +0000 UTC m=+1678.571592887" lastFinishedPulling="2025-11-29 00:27:45.860166239 +0000 UTC m=+1679.022059481" observedRunningTime="2025-11-29 00:27:46.462893565 +0000 UTC m=+1679.624786817" watchObservedRunningTime="2025-11-29 00:27:46.476959107 +0000 UTC m=+1679.638852349" Nov 29 00:27:49 crc kubenswrapper[4931]: I1129 00:27:49.048017 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-6dghl"] Nov 29 00:27:49 crc kubenswrapper[4931]: I1129 00:27:49.056343 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-6dghl"] Nov 29 00:27:49 crc kubenswrapper[4931]: I1129 00:27:49.228138 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3eb2f135-dcd7-4280-afa3-8f3fd1201b05" path="/var/lib/kubelet/pods/3eb2f135-dcd7-4280-afa3-8f3fd1201b05/volumes" Nov 29 00:27:49 crc kubenswrapper[4931]: I1129 00:27:49.473256 4931 generic.go:334] "Generic (PLEG): container finished" podID="2921aa97-795f-434a-ac6e-009e6555e864" containerID="a771a225b7abc6155350e7c6ccf53c7cf1ca505d8a17f1acd5249bd91879c1c5" exitCode=0 Nov 29 00:27:49 crc kubenswrapper[4931]: I1129 00:27:49.473303 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" event={"ID":"2921aa97-795f-434a-ac6e-009e6555e864","Type":"ContainerDied","Data":"a771a225b7abc6155350e7c6ccf53c7cf1ca505d8a17f1acd5249bd91879c1c5"} Nov 29 00:27:50 crc kubenswrapper[4931]: I1129 00:27:50.030278 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-9f3d-account-create-update-rfwpv"] Nov 29 00:27:50 crc kubenswrapper[4931]: I1129 00:27:50.046415 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-9f3d-account-create-update-rfwpv"] Nov 29 00:27:50 crc kubenswrapper[4931]: I1129 00:27:50.869278 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:50 crc kubenswrapper[4931]: I1129 00:27:50.973204 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2921aa97-795f-434a-ac6e-009e6555e864-ssh-key\") pod \"2921aa97-795f-434a-ac6e-009e6555e864\" (UID: \"2921aa97-795f-434a-ac6e-009e6555e864\") " Nov 29 00:27:50 crc kubenswrapper[4931]: I1129 00:27:50.973244 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22l6w\" (UniqueName: \"kubernetes.io/projected/2921aa97-795f-434a-ac6e-009e6555e864-kube-api-access-22l6w\") pod \"2921aa97-795f-434a-ac6e-009e6555e864\" (UID: \"2921aa97-795f-434a-ac6e-009e6555e864\") " Nov 29 00:27:50 crc kubenswrapper[4931]: I1129 00:27:50.973304 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2921aa97-795f-434a-ac6e-009e6555e864-inventory\") pod \"2921aa97-795f-434a-ac6e-009e6555e864\" (UID: \"2921aa97-795f-434a-ac6e-009e6555e864\") " Nov 29 00:27:50 crc kubenswrapper[4931]: I1129 00:27:50.979015 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2921aa97-795f-434a-ac6e-009e6555e864-kube-api-access-22l6w" (OuterVolumeSpecName: "kube-api-access-22l6w") pod "2921aa97-795f-434a-ac6e-009e6555e864" (UID: "2921aa97-795f-434a-ac6e-009e6555e864"). InnerVolumeSpecName "kube-api-access-22l6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.001084 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2921aa97-795f-434a-ac6e-009e6555e864-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2921aa97-795f-434a-ac6e-009e6555e864" (UID: "2921aa97-795f-434a-ac6e-009e6555e864"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.001481 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2921aa97-795f-434a-ac6e-009e6555e864-inventory" (OuterVolumeSpecName: "inventory") pod "2921aa97-795f-434a-ac6e-009e6555e864" (UID: "2921aa97-795f-434a-ac6e-009e6555e864"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.075869 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2921aa97-795f-434a-ac6e-009e6555e864-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.075917 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22l6w\" (UniqueName: \"kubernetes.io/projected/2921aa97-795f-434a-ac6e-009e6555e864-kube-api-access-22l6w\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.075935 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2921aa97-795f-434a-ac6e-009e6555e864-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.237454 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="269a3ba4-671b-47fd-8d81-960cc79810cb" path="/var/lib/kubelet/pods/269a3ba4-671b-47fd-8d81-960cc79810cb/volumes" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.499015 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" event={"ID":"2921aa97-795f-434a-ac6e-009e6555e864","Type":"ContainerDied","Data":"13a5ff8d3071f36dfc4f102fce50ca1b3f356983c1771b35ef49c39b1275003d"} Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.499379 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13a5ff8d3071f36dfc4f102fce50ca1b3f356983c1771b35ef49c39b1275003d" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.499094 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-n6rmv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.624368 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv"] Nov 29 00:27:51 crc kubenswrapper[4931]: E1129 00:27:51.624826 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2921aa97-795f-434a-ac6e-009e6555e864" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.624846 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="2921aa97-795f-434a-ac6e-009e6555e864" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.625103 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="2921aa97-795f-434a-ac6e-009e6555e864" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.625954 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.629024 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.629082 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.629497 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.629724 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.643900 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv"] Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.795034 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.795167 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.795219 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7htqg\" (UniqueName: \"kubernetes.io/projected/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-kube-api-access-7htqg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.795329 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.897534 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.897635 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.897673 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7htqg\" (UniqueName: \"kubernetes.io/projected/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-kube-api-access-7htqg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.897753 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.902727 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.904152 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.904257 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.921077 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7htqg\" (UniqueName: \"kubernetes.io/projected/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-kube-api-access-7htqg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:51 crc kubenswrapper[4931]: I1129 00:27:51.950017 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:27:52 crc kubenswrapper[4931]: I1129 00:27:52.501551 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv"] Nov 29 00:27:53 crc kubenswrapper[4931]: I1129 00:27:53.527454 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" event={"ID":"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71","Type":"ContainerStarted","Data":"0ef6ae455544d85498b2b05505e67fcb965d13e5abf822c696fc5997b35f653f"} Nov 29 00:27:53 crc kubenswrapper[4931]: I1129 00:27:53.528092 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" event={"ID":"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71","Type":"ContainerStarted","Data":"be9cf4162b18740121772ff3bb7b6d3b4717631e0683ea00eec35a8daa8f892e"} Nov 29 00:27:53 crc kubenswrapper[4931]: I1129 00:27:53.568716 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" podStartSLOduration=2.119960208 podStartE2EDuration="2.568693266s" podCreationTimestamp="2025-11-29 00:27:51 +0000 UTC" firstStartedPulling="2025-11-29 00:27:52.516878357 +0000 UTC m=+1685.678771599" lastFinishedPulling="2025-11-29 00:27:52.965611425 +0000 UTC m=+1686.127504657" observedRunningTime="2025-11-29 00:27:53.554964081 +0000 UTC m=+1686.716857373" watchObservedRunningTime="2025-11-29 00:27:53.568693266 +0000 UTC m=+1686.730586508" Nov 29 00:27:53 crc kubenswrapper[4931]: I1129 00:27:53.735005 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:27:53 crc kubenswrapper[4931]: I1129 00:27:53.735111 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:27:53 crc kubenswrapper[4931]: I1129 00:27:53.735189 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:27:53 crc kubenswrapper[4931]: I1129 00:27:53.736696 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 00:27:53 crc kubenswrapper[4931]: I1129 00:27:53.736871 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" gracePeriod=600 Nov 29 00:27:53 crc kubenswrapper[4931]: E1129 00:27:53.870846 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:27:54 crc kubenswrapper[4931]: I1129 00:27:54.546168 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" exitCode=0 Nov 29 00:27:54 crc kubenswrapper[4931]: I1129 00:27:54.546271 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357"} Nov 29 00:27:54 crc kubenswrapper[4931]: I1129 00:27:54.547083 4931 scope.go:117] "RemoveContainer" containerID="e130783d6b1a728f7dced4948fd5956d2a8b18dc3e74e1cf5f76c0244465eeba" Nov 29 00:27:54 crc kubenswrapper[4931]: I1129 00:27:54.549087 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:27:54 crc kubenswrapper[4931]: E1129 00:27:54.549534 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:27:55 crc kubenswrapper[4931]: I1129 00:27:55.048594 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ed60-account-create-update-9thjn"] Nov 29 00:27:55 crc kubenswrapper[4931]: I1129 00:27:55.063847 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-ed60-account-create-update-9thjn"] Nov 29 00:27:55 crc kubenswrapper[4931]: I1129 00:27:55.232941 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f55707a4-27e4-496f-ae85-1e07ddf91a5b" path="/var/lib/kubelet/pods/f55707a4-27e4-496f-ae85-1e07ddf91a5b/volumes" Nov 29 00:27:56 crc kubenswrapper[4931]: I1129 00:27:56.043559 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-ab0a-account-create-update-cjkxg"] Nov 29 00:27:56 crc kubenswrapper[4931]: I1129 00:27:56.055920 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-n4xfr"] Nov 29 00:27:56 crc kubenswrapper[4931]: I1129 00:27:56.077934 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-n4xfr"] Nov 29 00:27:56 crc kubenswrapper[4931]: I1129 00:27:56.078012 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-ab0a-account-create-update-cjkxg"] Nov 29 00:27:56 crc kubenswrapper[4931]: I1129 00:27:56.079422 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-f7mvr"] Nov 29 00:27:56 crc kubenswrapper[4931]: I1129 00:27:56.085864 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-f7mvr"] Nov 29 00:27:57 crc kubenswrapper[4931]: I1129 00:27:57.233159 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0032d91c-6fc0-4f0a-8410-ccf56967c6f8" path="/var/lib/kubelet/pods/0032d91c-6fc0-4f0a-8410-ccf56967c6f8/volumes" Nov 29 00:27:57 crc kubenswrapper[4931]: I1129 00:27:57.234614 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e" path="/var/lib/kubelet/pods/a7c9f10f-3823-4af5-9f94-ee8c5ef79d2e/volumes" Nov 29 00:27:57 crc kubenswrapper[4931]: I1129 00:27:57.235881 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf" path="/var/lib/kubelet/pods/d7a6941c-cc1c-4b2b-89f4-b6b866ef2abf/volumes" Nov 29 00:28:06 crc kubenswrapper[4931]: I1129 00:28:06.213764 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:28:06 crc kubenswrapper[4931]: E1129 00:28:06.215225 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:28:14 crc kubenswrapper[4931]: I1129 00:28:14.045466 4931 scope.go:117] "RemoveContainer" containerID="226248938d875e035fa0361171b0b911e810a8bfa3d2c7e6365d02e15a8f3c63" Nov 29 00:28:14 crc kubenswrapper[4931]: I1129 00:28:14.095703 4931 scope.go:117] "RemoveContainer" containerID="03ae9252b88462a5986482396100bb757d37d0fc08ab3d06303c511433c81ce0" Nov 29 00:28:14 crc kubenswrapper[4931]: I1129 00:28:14.143765 4931 scope.go:117] "RemoveContainer" containerID="928440447fae8581dee16eaf9a28be17570e240cbdac22bfa7f79ce828c80dde" Nov 29 00:28:14 crc kubenswrapper[4931]: I1129 00:28:14.201396 4931 scope.go:117] "RemoveContainer" containerID="a8945aa397a1bede424f2dde6d339fb659a0295784a089857ee21da174e9ce76" Nov 29 00:28:14 crc kubenswrapper[4931]: I1129 00:28:14.226793 4931 scope.go:117] "RemoveContainer" containerID="da47ddcf5db4bd413ff292ed16e42b6fbdcea7fac0fd2eb7ef57aecf5b1f054e" Nov 29 00:28:14 crc kubenswrapper[4931]: I1129 00:28:14.279857 4931 scope.go:117] "RemoveContainer" containerID="e7d2b957ede388e8fe9b73b67afd2a627fc22df45105975921c1c7fe66434928" Nov 29 00:28:14 crc kubenswrapper[4931]: I1129 00:28:14.306308 4931 scope.go:117] "RemoveContainer" containerID="28691ee7f4e971382326216c344002bbde7ae7ff76122bcb79215379b553dedf" Nov 29 00:28:14 crc kubenswrapper[4931]: I1129 00:28:14.331678 4931 scope.go:117] "RemoveContainer" containerID="fe8cf2636fb6ca05950e7503341d657922cff8234e607cd782332052fe55a21b" Nov 29 00:28:14 crc kubenswrapper[4931]: I1129 00:28:14.357446 4931 scope.go:117] "RemoveContainer" containerID="0af4b3337efba61d068d58053b627c4a625de291fd684680f1e0896589628714" Nov 29 00:28:14 crc kubenswrapper[4931]: I1129 00:28:14.379502 4931 scope.go:117] "RemoveContainer" containerID="ecaac952158742287c253bb03fdfe5e9c8c2b43a4749cd15a510cd6aac3d42ca" Nov 29 00:28:14 crc kubenswrapper[4931]: I1129 00:28:14.454621 4931 scope.go:117] "RemoveContainer" containerID="5cd01f9aec53fc4d19fcaeff2d34dd831adf986bd86afa31dd5492e48d08ff66" Nov 29 00:28:20 crc kubenswrapper[4931]: I1129 00:28:20.212628 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:28:20 crc kubenswrapper[4931]: E1129 00:28:20.213767 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:28:22 crc kubenswrapper[4931]: I1129 00:28:22.043699 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-d4wgk"] Nov 29 00:28:22 crc kubenswrapper[4931]: I1129 00:28:22.053882 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-rrpld"] Nov 29 00:28:22 crc kubenswrapper[4931]: I1129 00:28:22.062797 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-86a9-account-create-update-n4ljg"] Nov 29 00:28:22 crc kubenswrapper[4931]: I1129 00:28:22.074134 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-0c9c-account-create-update-6bwg6"] Nov 29 00:28:22 crc kubenswrapper[4931]: I1129 00:28:22.081486 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-d4wgk"] Nov 29 00:28:22 crc kubenswrapper[4931]: I1129 00:28:22.088188 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-2d72-account-create-update-qb5cv"] Nov 29 00:28:22 crc kubenswrapper[4931]: I1129 00:28:22.095591 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-kvkrd"] Nov 29 00:28:22 crc kubenswrapper[4931]: I1129 00:28:22.102738 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-86a9-account-create-update-n4ljg"] Nov 29 00:28:22 crc kubenswrapper[4931]: I1129 00:28:22.110444 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-rrpld"] Nov 29 00:28:22 crc kubenswrapper[4931]: I1129 00:28:22.117995 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-0c9c-account-create-update-6bwg6"] Nov 29 00:28:22 crc kubenswrapper[4931]: I1129 00:28:22.125161 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-2d72-account-create-update-qb5cv"] Nov 29 00:28:22 crc kubenswrapper[4931]: I1129 00:28:22.133554 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-kvkrd"] Nov 29 00:28:23 crc kubenswrapper[4931]: I1129 00:28:23.223035 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d367827-344f-49b6-9b67-9e7b08d2c89f" path="/var/lib/kubelet/pods/1d367827-344f-49b6-9b67-9e7b08d2c89f/volumes" Nov 29 00:28:23 crc kubenswrapper[4931]: I1129 00:28:23.224065 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d93d624-611c-405b-925c-41a3e159183d" path="/var/lib/kubelet/pods/2d93d624-611c-405b-925c-41a3e159183d/volumes" Nov 29 00:28:23 crc kubenswrapper[4931]: I1129 00:28:23.224691 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="897f5cd4-7cba-452f-bf48-ebe6c3333b76" path="/var/lib/kubelet/pods/897f5cd4-7cba-452f-bf48-ebe6c3333b76/volumes" Nov 29 00:28:23 crc kubenswrapper[4931]: I1129 00:28:23.225395 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92aad38f-3527-480c-ac71-23043a012568" path="/var/lib/kubelet/pods/92aad38f-3527-480c-ac71-23043a012568/volumes" Nov 29 00:28:23 crc kubenswrapper[4931]: I1129 00:28:23.226653 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abb709fd-4b15-4b57-bebd-2e1f945c4334" path="/var/lib/kubelet/pods/abb709fd-4b15-4b57-bebd-2e1f945c4334/volumes" Nov 29 00:28:23 crc kubenswrapper[4931]: I1129 00:28:23.227183 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3a60c30-b568-4927-bbe0-fcebead290cb" path="/var/lib/kubelet/pods/e3a60c30-b568-4927-bbe0-fcebead290cb/volumes" Nov 29 00:28:32 crc kubenswrapper[4931]: I1129 00:28:32.052917 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-ltd4m"] Nov 29 00:28:32 crc kubenswrapper[4931]: I1129 00:28:32.060597 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-ltd4m"] Nov 29 00:28:32 crc kubenswrapper[4931]: I1129 00:28:32.212207 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:28:32 crc kubenswrapper[4931]: E1129 00:28:32.212491 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:28:33 crc kubenswrapper[4931]: I1129 00:28:33.232873 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fae16d0f-aa7b-4f9b-aff2-896708ff5061" path="/var/lib/kubelet/pods/fae16d0f-aa7b-4f9b-aff2-896708ff5061/volumes" Nov 29 00:28:44 crc kubenswrapper[4931]: I1129 00:28:44.213998 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:28:44 crc kubenswrapper[4931]: E1129 00:28:44.215072 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:28:46 crc kubenswrapper[4931]: I1129 00:28:46.047706 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-dj2pb"] Nov 29 00:28:46 crc kubenswrapper[4931]: I1129 00:28:46.061727 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-dj2pb"] Nov 29 00:28:47 crc kubenswrapper[4931]: I1129 00:28:47.234664 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c957f2c-45cc-4a11-8019-e20de202266b" path="/var/lib/kubelet/pods/2c957f2c-45cc-4a11-8019-e20de202266b/volumes" Nov 29 00:28:56 crc kubenswrapper[4931]: I1129 00:28:56.212847 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:28:56 crc kubenswrapper[4931]: E1129 00:28:56.213790 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:29:09 crc kubenswrapper[4931]: I1129 00:29:09.212874 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:29:09 crc kubenswrapper[4931]: E1129 00:29:09.213759 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:29:14 crc kubenswrapper[4931]: I1129 00:29:14.061941 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-26gvb"] Nov 29 00:29:14 crc kubenswrapper[4931]: I1129 00:29:14.075409 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-26gvb"] Nov 29 00:29:14 crc kubenswrapper[4931]: I1129 00:29:14.691071 4931 scope.go:117] "RemoveContainer" containerID="23bcf002d1c5753522b14b9cebb06be0abf86762a0958c6bc9a4a018808cd16b" Nov 29 00:29:14 crc kubenswrapper[4931]: I1129 00:29:14.721065 4931 scope.go:117] "RemoveContainer" containerID="01d7fd77664235a8f6c4e5799750b304ad66c6e43d7e9dda8d763f02f0a460c1" Nov 29 00:29:14 crc kubenswrapper[4931]: I1129 00:29:14.789769 4931 scope.go:117] "RemoveContainer" containerID="e0fd7cff7473147e00ffdc10c155c7d1bc55e6e49e7189c3ef409d989ab1d86a" Nov 29 00:29:14 crc kubenswrapper[4931]: I1129 00:29:14.822731 4931 scope.go:117] "RemoveContainer" containerID="549aecf595ecff89023b1a2b77d9b714fa543e1a33c0e1a53ad670b5080a7092" Nov 29 00:29:14 crc kubenswrapper[4931]: I1129 00:29:14.888436 4931 scope.go:117] "RemoveContainer" containerID="46a7a92d8f936282a47882ed62369636caac887e41020f2f68642a95cd156509" Nov 29 00:29:14 crc kubenswrapper[4931]: I1129 00:29:14.910605 4931 scope.go:117] "RemoveContainer" containerID="6f88355e1883ba361cc10dfbeeb09a2bce98d2f2d3f6e2d4108ff6798c15607e" Nov 29 00:29:14 crc kubenswrapper[4931]: I1129 00:29:14.953192 4931 scope.go:117] "RemoveContainer" containerID="d25680c15e3bf4b84c3c75307bee7854603df02a9c964738af8672da521e16e3" Nov 29 00:29:14 crc kubenswrapper[4931]: I1129 00:29:14.988708 4931 scope.go:117] "RemoveContainer" containerID="0c5009dbbe024a398e932da23bd1e88284cd87ee64660cab3a23aacacd9f7db3" Nov 29 00:29:15 crc kubenswrapper[4931]: I1129 00:29:15.227423 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a72a028-d521-466b-9924-eba4595e3750" path="/var/lib/kubelet/pods/2a72a028-d521-466b-9924-eba4595e3750/volumes" Nov 29 00:29:20 crc kubenswrapper[4931]: I1129 00:29:20.213139 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:29:20 crc kubenswrapper[4931]: E1129 00:29:20.214280 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:29:23 crc kubenswrapper[4931]: I1129 00:29:23.043714 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-wknnq"] Nov 29 00:29:23 crc kubenswrapper[4931]: I1129 00:29:23.055457 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-z2cs6"] Nov 29 00:29:23 crc kubenswrapper[4931]: I1129 00:29:23.066527 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-wknnq"] Nov 29 00:29:23 crc kubenswrapper[4931]: I1129 00:29:23.072868 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-z2cs6"] Nov 29 00:29:23 crc kubenswrapper[4931]: I1129 00:29:23.222947 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6041acf3-4620-48a5-8455-6c8cdb50cb0b" path="/var/lib/kubelet/pods/6041acf3-4620-48a5-8455-6c8cdb50cb0b/volumes" Nov 29 00:29:23 crc kubenswrapper[4931]: I1129 00:29:23.223681 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e6ded52-583f-44e7-853c-30c354886ae9" path="/var/lib/kubelet/pods/8e6ded52-583f-44e7-853c-30c354886ae9/volumes" Nov 29 00:29:27 crc kubenswrapper[4931]: I1129 00:29:27.027893 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-pml9v"] Nov 29 00:29:27 crc kubenswrapper[4931]: I1129 00:29:27.038670 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-pml9v"] Nov 29 00:29:27 crc kubenswrapper[4931]: I1129 00:29:27.231785 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27762870-383b-4e1b-8d28-e6e7c96d635a" path="/var/lib/kubelet/pods/27762870-383b-4e1b-8d28-e6e7c96d635a/volumes" Nov 29 00:29:29 crc kubenswrapper[4931]: I1129 00:29:29.046475 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-dn48v"] Nov 29 00:29:29 crc kubenswrapper[4931]: I1129 00:29:29.056147 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-dn48v"] Nov 29 00:29:29 crc kubenswrapper[4931]: I1129 00:29:29.232393 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02ce26c3-6755-479e-93c4-295ef8a3ba2a" path="/var/lib/kubelet/pods/02ce26c3-6755-479e-93c4-295ef8a3ba2a/volumes" Nov 29 00:29:32 crc kubenswrapper[4931]: I1129 00:29:32.213342 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:29:32 crc kubenswrapper[4931]: E1129 00:29:32.214136 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:29:44 crc kubenswrapper[4931]: I1129 00:29:44.213141 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:29:44 crc kubenswrapper[4931]: E1129 00:29:44.216174 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:29:59 crc kubenswrapper[4931]: I1129 00:29:59.213235 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:29:59 crc kubenswrapper[4931]: E1129 00:29:59.214176 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.160110 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8"] Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.161593 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.163671 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.165060 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.165059 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24jnr\" (UniqueName: \"kubernetes.io/projected/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-kube-api-access-24jnr\") pod \"collect-profiles-29406270-l8vn8\" (UID: \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.165194 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-config-volume\") pod \"collect-profiles-29406270-l8vn8\" (UID: \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.165430 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-secret-volume\") pod \"collect-profiles-29406270-l8vn8\" (UID: \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.196550 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8"] Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.268036 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-secret-volume\") pod \"collect-profiles-29406270-l8vn8\" (UID: \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.269391 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24jnr\" (UniqueName: \"kubernetes.io/projected/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-kube-api-access-24jnr\") pod \"collect-profiles-29406270-l8vn8\" (UID: \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.269714 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-config-volume\") pod \"collect-profiles-29406270-l8vn8\" (UID: \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.271461 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-config-volume\") pod \"collect-profiles-29406270-l8vn8\" (UID: \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.277561 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-secret-volume\") pod \"collect-profiles-29406270-l8vn8\" (UID: \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.296143 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24jnr\" (UniqueName: \"kubernetes.io/projected/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-kube-api-access-24jnr\") pod \"collect-profiles-29406270-l8vn8\" (UID: \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.492494 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:00 crc kubenswrapper[4931]: I1129 00:30:00.971680 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8"] Nov 29 00:30:01 crc kubenswrapper[4931]: I1129 00:30:01.914732 4931 generic.go:334] "Generic (PLEG): container finished" podID="489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6" containerID="7884d5457e9b5abc28e60b8ed77e2eb30fb101add86ac5f94b1002c5faf2c62a" exitCode=0 Nov 29 00:30:01 crc kubenswrapper[4931]: I1129 00:30:01.914832 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" event={"ID":"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6","Type":"ContainerDied","Data":"7884d5457e9b5abc28e60b8ed77e2eb30fb101add86ac5f94b1002c5faf2c62a"} Nov 29 00:30:01 crc kubenswrapper[4931]: I1129 00:30:01.914889 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" event={"ID":"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6","Type":"ContainerStarted","Data":"432146324b8bafd5031aa59db64cf6274971d6f881c9ff195d72e896ceae93d3"} Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.309093 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.333647 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24jnr\" (UniqueName: \"kubernetes.io/projected/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-kube-api-access-24jnr\") pod \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\" (UID: \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\") " Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.333715 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-config-volume\") pod \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\" (UID: \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\") " Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.333989 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-secret-volume\") pod \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\" (UID: \"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6\") " Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.335630 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-config-volume" (OuterVolumeSpecName: "config-volume") pod "489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6" (UID: "489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.339689 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-kube-api-access-24jnr" (OuterVolumeSpecName: "kube-api-access-24jnr") pod "489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6" (UID: "489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6"). InnerVolumeSpecName "kube-api-access-24jnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.339800 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6" (UID: "489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.436403 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24jnr\" (UniqueName: \"kubernetes.io/projected/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-kube-api-access-24jnr\") on node \"crc\" DevicePath \"\"" Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.436443 4931 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.436454 4931 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.944140 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" event={"ID":"489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6","Type":"ContainerDied","Data":"432146324b8bafd5031aa59db64cf6274971d6f881c9ff195d72e896ceae93d3"} Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.944210 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="432146324b8bafd5031aa59db64cf6274971d6f881c9ff195d72e896ceae93d3" Nov 29 00:30:03 crc kubenswrapper[4931]: I1129 00:30:03.944308 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8" Nov 29 00:30:10 crc kubenswrapper[4931]: I1129 00:30:10.212579 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:30:10 crc kubenswrapper[4931]: E1129 00:30:10.213770 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:30:15 crc kubenswrapper[4931]: I1129 00:30:15.161724 4931 scope.go:117] "RemoveContainer" containerID="3eb05edd39e3fe12742b69fd09c8953dee41262283efea74b3bd4418cca0f085" Nov 29 00:30:15 crc kubenswrapper[4931]: I1129 00:30:15.227239 4931 scope.go:117] "RemoveContainer" containerID="f1ad50d40ba416d98d59313fc595dc27e2284b8f1c20fc2459e688f103bfe16d" Nov 29 00:30:15 crc kubenswrapper[4931]: I1129 00:30:15.293252 4931 scope.go:117] "RemoveContainer" containerID="9a39ab75b839a4c6e4d50a13762239245f3d69a686155574f10f887436ee76e3" Nov 29 00:30:15 crc kubenswrapper[4931]: I1129 00:30:15.321238 4931 scope.go:117] "RemoveContainer" containerID="691315ec4d054a167417bbadfd4ddb939e88c1eed51bf22e2853223ea252e573" Nov 29 00:30:15 crc kubenswrapper[4931]: I1129 00:30:15.364585 4931 scope.go:117] "RemoveContainer" containerID="290aa1bf1bb68db7495e27dd7559d5ed0fdf5c812761bb0c24d52ba4dcb49598" Nov 29 00:30:23 crc kubenswrapper[4931]: I1129 00:30:23.211941 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:30:23 crc kubenswrapper[4931]: E1129 00:30:23.212589 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:30:32 crc kubenswrapper[4931]: I1129 00:30:32.050151 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-dw4x9"] Nov 29 00:30:32 crc kubenswrapper[4931]: I1129 00:30:32.064604 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-dw4x9"] Nov 29 00:30:32 crc kubenswrapper[4931]: I1129 00:30:32.077444 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-n899b"] Nov 29 00:30:32 crc kubenswrapper[4931]: I1129 00:30:32.090934 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-r7n7s"] Nov 29 00:30:32 crc kubenswrapper[4931]: I1129 00:30:32.098699 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-r7n7s"] Nov 29 00:30:32 crc kubenswrapper[4931]: I1129 00:30:32.107892 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-n899b"] Nov 29 00:30:33 crc kubenswrapper[4931]: I1129 00:30:33.056957 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-3603-account-create-update-8rgkv"] Nov 29 00:30:33 crc kubenswrapper[4931]: I1129 00:30:33.070886 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-4b32-account-create-update-nhrbj"] Nov 29 00:30:33 crc kubenswrapper[4931]: I1129 00:30:33.098179 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-3603-account-create-update-8rgkv"] Nov 29 00:30:33 crc kubenswrapper[4931]: I1129 00:30:33.108454 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-4b32-account-create-update-nhrbj"] Nov 29 00:30:33 crc kubenswrapper[4931]: I1129 00:30:33.115722 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-68b1-account-create-update-4flsl"] Nov 29 00:30:33 crc kubenswrapper[4931]: I1129 00:30:33.124638 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-68b1-account-create-update-4flsl"] Nov 29 00:30:33 crc kubenswrapper[4931]: I1129 00:30:33.235447 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f6ca58a-4bc4-4277-891d-db214595cb56" path="/var/lib/kubelet/pods/2f6ca58a-4bc4-4277-891d-db214595cb56/volumes" Nov 29 00:30:33 crc kubenswrapper[4931]: I1129 00:30:33.236402 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46687ab2-074a-4007-b89a-a962c9b1d14d" path="/var/lib/kubelet/pods/46687ab2-074a-4007-b89a-a962c9b1d14d/volumes" Nov 29 00:30:33 crc kubenswrapper[4931]: I1129 00:30:33.237161 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f82d911-133a-4c7a-87a4-9684bb066cc0" path="/var/lib/kubelet/pods/5f82d911-133a-4c7a-87a4-9684bb066cc0/volumes" Nov 29 00:30:33 crc kubenswrapper[4931]: I1129 00:30:33.237866 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6284f70c-aaa2-4182-98fd-82fb59295e1f" path="/var/lib/kubelet/pods/6284f70c-aaa2-4182-98fd-82fb59295e1f/volumes" Nov 29 00:30:33 crc kubenswrapper[4931]: I1129 00:30:33.239221 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5" path="/var/lib/kubelet/pods/a52f3c1d-5aa8-4c5c-a51b-71b29ea5e0f5/volumes" Nov 29 00:30:33 crc kubenswrapper[4931]: I1129 00:30:33.239976 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a57e64d1-340e-4e80-8112-16625e6d7d6e" path="/var/lib/kubelet/pods/a57e64d1-340e-4e80-8112-16625e6d7d6e/volumes" Nov 29 00:30:37 crc kubenswrapper[4931]: I1129 00:30:37.214292 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:30:37 crc kubenswrapper[4931]: E1129 00:30:37.215429 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:30:48 crc kubenswrapper[4931]: I1129 00:30:48.212595 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:30:48 crc kubenswrapper[4931]: E1129 00:30:48.213375 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:31:03 crc kubenswrapper[4931]: I1129 00:31:03.055816 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b4dlr"] Nov 29 00:31:03 crc kubenswrapper[4931]: I1129 00:31:03.063922 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b4dlr"] Nov 29 00:31:03 crc kubenswrapper[4931]: I1129 00:31:03.212305 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:31:03 crc kubenswrapper[4931]: E1129 00:31:03.212561 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:31:03 crc kubenswrapper[4931]: I1129 00:31:03.221450 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f40ca0f5-3fa9-4003-87cb-aed2feeaf545" path="/var/lib/kubelet/pods/f40ca0f5-3fa9-4003-87cb-aed2feeaf545/volumes" Nov 29 00:31:14 crc kubenswrapper[4931]: I1129 00:31:14.213346 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:31:14 crc kubenswrapper[4931]: E1129 00:31:14.214723 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:31:15 crc kubenswrapper[4931]: I1129 00:31:15.534537 4931 scope.go:117] "RemoveContainer" containerID="69bf7ea057acdfc9d180e8d17d962cc2b048ebda75fe1998c51db15f04dce044" Nov 29 00:31:15 crc kubenswrapper[4931]: I1129 00:31:15.567770 4931 scope.go:117] "RemoveContainer" containerID="69a758d0710ba7c61b3fa8ff23863c7140f3896b1f2832cdf1af16162b8bc156" Nov 29 00:31:15 crc kubenswrapper[4931]: I1129 00:31:15.642212 4931 scope.go:117] "RemoveContainer" containerID="d50ca037edd2752ec5e0926af44bb5d8cead4abc1e1ad1a96b51f1b117b215d1" Nov 29 00:31:15 crc kubenswrapper[4931]: I1129 00:31:15.662215 4931 scope.go:117] "RemoveContainer" containerID="d89c36e7e0927e0ca927465aaa61f9848057756bee2cd62e2c72bc146aa7e137" Nov 29 00:31:15 crc kubenswrapper[4931]: I1129 00:31:15.722912 4931 scope.go:117] "RemoveContainer" containerID="0489f9f59e6d194d1f2a44f58a42656cb27237e5e71d01d367eba8c79ebc2c70" Nov 29 00:31:15 crc kubenswrapper[4931]: I1129 00:31:15.780669 4931 scope.go:117] "RemoveContainer" containerID="a61489bacc4a02782d504b27decc932ac85e50c8676418a8d3d2613218a52f0d" Nov 29 00:31:15 crc kubenswrapper[4931]: I1129 00:31:15.828310 4931 scope.go:117] "RemoveContainer" containerID="f8eda5ccdc2aede73e4bc9eaab88cdd9f2ca4a37c25f35e76630e432175ea07b" Nov 29 00:31:19 crc kubenswrapper[4931]: I1129 00:31:19.760208 4931 generic.go:334] "Generic (PLEG): container finished" podID="6a0a53a8-99d7-4a99-bdf0-67f72b8cca71" containerID="0ef6ae455544d85498b2b05505e67fcb965d13e5abf822c696fc5997b35f653f" exitCode=0 Nov 29 00:31:19 crc kubenswrapper[4931]: I1129 00:31:19.760269 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" event={"ID":"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71","Type":"ContainerDied","Data":"0ef6ae455544d85498b2b05505e67fcb965d13e5abf822c696fc5997b35f653f"} Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.217157 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.284718 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-bootstrap-combined-ca-bundle\") pod \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.284893 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-ssh-key\") pod \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.285049 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7htqg\" (UniqueName: \"kubernetes.io/projected/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-kube-api-access-7htqg\") pod \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.285131 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-inventory\") pod \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\" (UID: \"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71\") " Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.306565 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-kube-api-access-7htqg" (OuterVolumeSpecName: "kube-api-access-7htqg") pod "6a0a53a8-99d7-4a99-bdf0-67f72b8cca71" (UID: "6a0a53a8-99d7-4a99-bdf0-67f72b8cca71"). InnerVolumeSpecName "kube-api-access-7htqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.309016 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6a0a53a8-99d7-4a99-bdf0-67f72b8cca71" (UID: "6a0a53a8-99d7-4a99-bdf0-67f72b8cca71"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.315477 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-inventory" (OuterVolumeSpecName: "inventory") pod "6a0a53a8-99d7-4a99-bdf0-67f72b8cca71" (UID: "6a0a53a8-99d7-4a99-bdf0-67f72b8cca71"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.317101 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6a0a53a8-99d7-4a99-bdf0-67f72b8cca71" (UID: "6a0a53a8-99d7-4a99-bdf0-67f72b8cca71"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.387335 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7htqg\" (UniqueName: \"kubernetes.io/projected/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-kube-api-access-7htqg\") on node \"crc\" DevicePath \"\"" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.387377 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.387393 4931 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.387404 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a0a53a8-99d7-4a99-bdf0-67f72b8cca71-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.783402 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" event={"ID":"6a0a53a8-99d7-4a99-bdf0-67f72b8cca71","Type":"ContainerDied","Data":"be9cf4162b18740121772ff3bb7b6d3b4717631e0683ea00eec35a8daa8f892e"} Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.783456 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.783464 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be9cf4162b18740121772ff3bb7b6d3b4717631e0683ea00eec35a8daa8f892e" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.911268 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6"] Nov 29 00:31:21 crc kubenswrapper[4931]: E1129 00:31:21.912685 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a0a53a8-99d7-4a99-bdf0-67f72b8cca71" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.912722 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a0a53a8-99d7-4a99-bdf0-67f72b8cca71" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 29 00:31:21 crc kubenswrapper[4931]: E1129 00:31:21.912760 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6" containerName="collect-profiles" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.912771 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6" containerName="collect-profiles" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.913088 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a0a53a8-99d7-4a99-bdf0-67f72b8cca71" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.913131 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6" containerName="collect-profiles" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.914099 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.917781 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.918667 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.919067 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.919250 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.919499 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6"] Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.998717 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/935c6e60-45b7-4859-bbd0-b147fa600465-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6\" (UID: \"935c6e60-45b7-4859-bbd0-b147fa600465\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:31:21 crc kubenswrapper[4931]: I1129 00:31:21.999014 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/935c6e60-45b7-4859-bbd0-b147fa600465-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6\" (UID: \"935c6e60-45b7-4859-bbd0-b147fa600465\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:31:22 crc kubenswrapper[4931]: I1129 00:31:21.999386 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw2xx\" (UniqueName: \"kubernetes.io/projected/935c6e60-45b7-4859-bbd0-b147fa600465-kube-api-access-zw2xx\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6\" (UID: \"935c6e60-45b7-4859-bbd0-b147fa600465\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:31:22 crc kubenswrapper[4931]: I1129 00:31:22.100686 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/935c6e60-45b7-4859-bbd0-b147fa600465-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6\" (UID: \"935c6e60-45b7-4859-bbd0-b147fa600465\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:31:22 crc kubenswrapper[4931]: I1129 00:31:22.100785 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw2xx\" (UniqueName: \"kubernetes.io/projected/935c6e60-45b7-4859-bbd0-b147fa600465-kube-api-access-zw2xx\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6\" (UID: \"935c6e60-45b7-4859-bbd0-b147fa600465\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:31:22 crc kubenswrapper[4931]: I1129 00:31:22.100855 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/935c6e60-45b7-4859-bbd0-b147fa600465-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6\" (UID: \"935c6e60-45b7-4859-bbd0-b147fa600465\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:31:22 crc kubenswrapper[4931]: I1129 00:31:22.106039 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/935c6e60-45b7-4859-bbd0-b147fa600465-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6\" (UID: \"935c6e60-45b7-4859-bbd0-b147fa600465\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:31:22 crc kubenswrapper[4931]: I1129 00:31:22.106143 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/935c6e60-45b7-4859-bbd0-b147fa600465-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6\" (UID: \"935c6e60-45b7-4859-bbd0-b147fa600465\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:31:22 crc kubenswrapper[4931]: I1129 00:31:22.120661 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw2xx\" (UniqueName: \"kubernetes.io/projected/935c6e60-45b7-4859-bbd0-b147fa600465-kube-api-access-zw2xx\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6\" (UID: \"935c6e60-45b7-4859-bbd0-b147fa600465\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:31:22 crc kubenswrapper[4931]: I1129 00:31:22.245251 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:31:22 crc kubenswrapper[4931]: I1129 00:31:22.894772 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6"] Nov 29 00:31:22 crc kubenswrapper[4931]: W1129 00:31:22.903945 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod935c6e60_45b7_4859_bbd0_b147fa600465.slice/crio-6d49edcacc893676280d62d1875fc1382953b17867f916663b08a0f3ae8eba82 WatchSource:0}: Error finding container 6d49edcacc893676280d62d1875fc1382953b17867f916663b08a0f3ae8eba82: Status 404 returned error can't find the container with id 6d49edcacc893676280d62d1875fc1382953b17867f916663b08a0f3ae8eba82 Nov 29 00:31:23 crc kubenswrapper[4931]: I1129 00:31:23.800454 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" event={"ID":"935c6e60-45b7-4859-bbd0-b147fa600465","Type":"ContainerStarted","Data":"6d49edcacc893676280d62d1875fc1382953b17867f916663b08a0f3ae8eba82"} Nov 29 00:31:24 crc kubenswrapper[4931]: I1129 00:31:24.813120 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" event={"ID":"935c6e60-45b7-4859-bbd0-b147fa600465","Type":"ContainerStarted","Data":"f6947708df7249911bdbc91496c01f462f0a7e414bf1eefa9e4a043595bedfd1"} Nov 29 00:31:25 crc kubenswrapper[4931]: I1129 00:31:25.027132 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" podStartSLOduration=3.263297402 podStartE2EDuration="4.027106001s" podCreationTimestamp="2025-11-29 00:31:21 +0000 UTC" firstStartedPulling="2025-11-29 00:31:22.905499258 +0000 UTC m=+1896.067392490" lastFinishedPulling="2025-11-29 00:31:23.669307817 +0000 UTC m=+1896.831201089" observedRunningTime="2025-11-29 00:31:24.836685066 +0000 UTC m=+1897.998578338" watchObservedRunningTime="2025-11-29 00:31:25.027106001 +0000 UTC m=+1898.188999263" Nov 29 00:31:25 crc kubenswrapper[4931]: I1129 00:31:25.033723 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-gf7vk"] Nov 29 00:31:25 crc kubenswrapper[4931]: I1129 00:31:25.047591 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-gf7vk"] Nov 29 00:31:25 crc kubenswrapper[4931]: I1129 00:31:25.223499 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="578b8169-9015-4faa-833c-25d8f7d75d10" path="/var/lib/kubelet/pods/578b8169-9015-4faa-833c-25d8f7d75d10/volumes" Nov 29 00:31:27 crc kubenswrapper[4931]: I1129 00:31:27.228182 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:31:27 crc kubenswrapper[4931]: E1129 00:31:27.228826 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:31:28 crc kubenswrapper[4931]: I1129 00:31:28.035409 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rd7sl"] Nov 29 00:31:28 crc kubenswrapper[4931]: I1129 00:31:28.045544 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rd7sl"] Nov 29 00:31:29 crc kubenswrapper[4931]: I1129 00:31:29.223284 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ce314fe-040a-4d82-925d-d48c52a993dd" path="/var/lib/kubelet/pods/8ce314fe-040a-4d82-925d-d48c52a993dd/volumes" Nov 29 00:31:40 crc kubenswrapper[4931]: I1129 00:31:40.212965 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:31:40 crc kubenswrapper[4931]: E1129 00:31:40.215382 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:31:55 crc kubenswrapper[4931]: I1129 00:31:55.212172 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:31:55 crc kubenswrapper[4931]: E1129 00:31:55.213132 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:32:08 crc kubenswrapper[4931]: I1129 00:32:08.212470 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:32:08 crc kubenswrapper[4931]: E1129 00:32:08.213392 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:32:13 crc kubenswrapper[4931]: I1129 00:32:13.057135 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-bwdbx"] Nov 29 00:32:13 crc kubenswrapper[4931]: I1129 00:32:13.067629 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-bwdbx"] Nov 29 00:32:13 crc kubenswrapper[4931]: I1129 00:32:13.227124 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f619a0b-81e1-4881-abaf-643f69860a61" path="/var/lib/kubelet/pods/9f619a0b-81e1-4881-abaf-643f69860a61/volumes" Nov 29 00:32:16 crc kubenswrapper[4931]: I1129 00:32:16.020209 4931 scope.go:117] "RemoveContainer" containerID="c687e3ab7c36b06a8733eb3998199211c2735d22306e26ffb70a209f8a330428" Nov 29 00:32:16 crc kubenswrapper[4931]: I1129 00:32:16.080312 4931 scope.go:117] "RemoveContainer" containerID="ca7d2ede55474b3056f1012b12fe3e926f0e93770a1184197b2197d9a96cc0fc" Nov 29 00:32:16 crc kubenswrapper[4931]: I1129 00:32:16.126526 4931 scope.go:117] "RemoveContainer" containerID="c2ad958301b735a0e4f56dfed8e1511d8d5d02af42a97b9c58bf0ef334a1ff1e" Nov 29 00:32:23 crc kubenswrapper[4931]: I1129 00:32:23.213615 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:32:23 crc kubenswrapper[4931]: E1129 00:32:23.216140 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:32:37 crc kubenswrapper[4931]: I1129 00:32:37.212348 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:32:37 crc kubenswrapper[4931]: E1129 00:32:37.213492 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.003305 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4nv4g"] Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.005872 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.021949 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4nv4g"] Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.160768 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c47b36f-2cf1-44bb-91af-c16c42a84d22-utilities\") pod \"redhat-operators-4nv4g\" (UID: \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\") " pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.161220 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlssc\" (UniqueName: \"kubernetes.io/projected/1c47b36f-2cf1-44bb-91af-c16c42a84d22-kube-api-access-nlssc\") pod \"redhat-operators-4nv4g\" (UID: \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\") " pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.161293 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c47b36f-2cf1-44bb-91af-c16c42a84d22-catalog-content\") pod \"redhat-operators-4nv4g\" (UID: \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\") " pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.263029 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c47b36f-2cf1-44bb-91af-c16c42a84d22-utilities\") pod \"redhat-operators-4nv4g\" (UID: \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\") " pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.263483 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c47b36f-2cf1-44bb-91af-c16c42a84d22-utilities\") pod \"redhat-operators-4nv4g\" (UID: \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\") " pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.263491 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlssc\" (UniqueName: \"kubernetes.io/projected/1c47b36f-2cf1-44bb-91af-c16c42a84d22-kube-api-access-nlssc\") pod \"redhat-operators-4nv4g\" (UID: \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\") " pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.263602 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c47b36f-2cf1-44bb-91af-c16c42a84d22-catalog-content\") pod \"redhat-operators-4nv4g\" (UID: \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\") " pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.264267 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c47b36f-2cf1-44bb-91af-c16c42a84d22-catalog-content\") pod \"redhat-operators-4nv4g\" (UID: \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\") " pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.286663 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlssc\" (UniqueName: \"kubernetes.io/projected/1c47b36f-2cf1-44bb-91af-c16c42a84d22-kube-api-access-nlssc\") pod \"redhat-operators-4nv4g\" (UID: \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\") " pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.368028 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:44 crc kubenswrapper[4931]: I1129 00:32:44.805412 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4nv4g"] Nov 29 00:32:45 crc kubenswrapper[4931]: I1129 00:32:45.721536 4931 generic.go:334] "Generic (PLEG): container finished" podID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" containerID="64d68d2e05275ca0781abd42f038c6ac48dfcff06b16d6fc0dea82168a63676d" exitCode=0 Nov 29 00:32:45 crc kubenswrapper[4931]: I1129 00:32:45.721597 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nv4g" event={"ID":"1c47b36f-2cf1-44bb-91af-c16c42a84d22","Type":"ContainerDied","Data":"64d68d2e05275ca0781abd42f038c6ac48dfcff06b16d6fc0dea82168a63676d"} Nov 29 00:32:45 crc kubenswrapper[4931]: I1129 00:32:45.721876 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nv4g" event={"ID":"1c47b36f-2cf1-44bb-91af-c16c42a84d22","Type":"ContainerStarted","Data":"eaa626c4db61e2b618379379bac2efa38d08f56d86a4790068ec5208fb130ee0"} Nov 29 00:32:45 crc kubenswrapper[4931]: I1129 00:32:45.723657 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 00:32:46 crc kubenswrapper[4931]: I1129 00:32:46.736622 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nv4g" event={"ID":"1c47b36f-2cf1-44bb-91af-c16c42a84d22","Type":"ContainerStarted","Data":"d9c752137312be22ad9cd96fc9b25ccb983a1b5666ce8020a1265aceeb30de90"} Nov 29 00:32:48 crc kubenswrapper[4931]: I1129 00:32:48.212203 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:32:48 crc kubenswrapper[4931]: E1129 00:32:48.212734 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:32:49 crc kubenswrapper[4931]: I1129 00:32:49.771430 4931 generic.go:334] "Generic (PLEG): container finished" podID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" containerID="d9c752137312be22ad9cd96fc9b25ccb983a1b5666ce8020a1265aceeb30de90" exitCode=0 Nov 29 00:32:49 crc kubenswrapper[4931]: I1129 00:32:49.771466 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nv4g" event={"ID":"1c47b36f-2cf1-44bb-91af-c16c42a84d22","Type":"ContainerDied","Data":"d9c752137312be22ad9cd96fc9b25ccb983a1b5666ce8020a1265aceeb30de90"} Nov 29 00:32:50 crc kubenswrapper[4931]: I1129 00:32:50.781859 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nv4g" event={"ID":"1c47b36f-2cf1-44bb-91af-c16c42a84d22","Type":"ContainerStarted","Data":"109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da"} Nov 29 00:32:50 crc kubenswrapper[4931]: I1129 00:32:50.811390 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4nv4g" podStartSLOduration=3.207191063 podStartE2EDuration="7.811370302s" podCreationTimestamp="2025-11-29 00:32:43 +0000 UTC" firstStartedPulling="2025-11-29 00:32:45.723323677 +0000 UTC m=+1978.885216909" lastFinishedPulling="2025-11-29 00:32:50.327502916 +0000 UTC m=+1983.489396148" observedRunningTime="2025-11-29 00:32:50.806328629 +0000 UTC m=+1983.968221871" watchObservedRunningTime="2025-11-29 00:32:50.811370302 +0000 UTC m=+1983.973263544" Nov 29 00:32:54 crc kubenswrapper[4931]: I1129 00:32:54.368912 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:54 crc kubenswrapper[4931]: I1129 00:32:54.370164 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:32:55 crc kubenswrapper[4931]: I1129 00:32:55.412820 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4nv4g" podUID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" containerName="registry-server" probeResult="failure" output=< Nov 29 00:32:55 crc kubenswrapper[4931]: timeout: failed to connect service ":50051" within 1s Nov 29 00:32:55 crc kubenswrapper[4931]: > Nov 29 00:33:03 crc kubenswrapper[4931]: I1129 00:33:03.212696 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:33:03 crc kubenswrapper[4931]: I1129 00:33:03.922219 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"bc12cbef83d264cf45b46aedbbb0832fdfe3ed670622f291973be8df8b6edd8a"} Nov 29 00:33:04 crc kubenswrapper[4931]: I1129 00:33:04.476218 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:33:04 crc kubenswrapper[4931]: I1129 00:33:04.531939 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:33:04 crc kubenswrapper[4931]: I1129 00:33:04.726303 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4nv4g"] Nov 29 00:33:05 crc kubenswrapper[4931]: I1129 00:33:05.950371 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4nv4g" podUID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" containerName="registry-server" containerID="cri-o://109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da" gracePeriod=2 Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.464124 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.629317 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c47b36f-2cf1-44bb-91af-c16c42a84d22-catalog-content\") pod \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\" (UID: \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\") " Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.636286 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlssc\" (UniqueName: \"kubernetes.io/projected/1c47b36f-2cf1-44bb-91af-c16c42a84d22-kube-api-access-nlssc\") pod \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\" (UID: \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\") " Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.637190 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c47b36f-2cf1-44bb-91af-c16c42a84d22-utilities\") pod \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\" (UID: \"1c47b36f-2cf1-44bb-91af-c16c42a84d22\") " Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.638015 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c47b36f-2cf1-44bb-91af-c16c42a84d22-utilities" (OuterVolumeSpecName: "utilities") pod "1c47b36f-2cf1-44bb-91af-c16c42a84d22" (UID: "1c47b36f-2cf1-44bb-91af-c16c42a84d22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.644106 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c47b36f-2cf1-44bb-91af-c16c42a84d22-kube-api-access-nlssc" (OuterVolumeSpecName: "kube-api-access-nlssc") pod "1c47b36f-2cf1-44bb-91af-c16c42a84d22" (UID: "1c47b36f-2cf1-44bb-91af-c16c42a84d22"). InnerVolumeSpecName "kube-api-access-nlssc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.738986 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c47b36f-2cf1-44bb-91af-c16c42a84d22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c47b36f-2cf1-44bb-91af-c16c42a84d22" (UID: "1c47b36f-2cf1-44bb-91af-c16c42a84d22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.739677 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c47b36f-2cf1-44bb-91af-c16c42a84d22-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.739717 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c47b36f-2cf1-44bb-91af-c16c42a84d22-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.739731 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlssc\" (UniqueName: \"kubernetes.io/projected/1c47b36f-2cf1-44bb-91af-c16c42a84d22-kube-api-access-nlssc\") on node \"crc\" DevicePath \"\"" Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.965725 4931 generic.go:334] "Generic (PLEG): container finished" podID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" containerID="109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da" exitCode=0 Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.965770 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4nv4g" Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.965800 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nv4g" event={"ID":"1c47b36f-2cf1-44bb-91af-c16c42a84d22","Type":"ContainerDied","Data":"109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da"} Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.965883 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nv4g" event={"ID":"1c47b36f-2cf1-44bb-91af-c16c42a84d22","Type":"ContainerDied","Data":"eaa626c4db61e2b618379379bac2efa38d08f56d86a4790068ec5208fb130ee0"} Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.965911 4931 scope.go:117] "RemoveContainer" containerID="109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da" Nov 29 00:33:06 crc kubenswrapper[4931]: I1129 00:33:06.997857 4931 scope.go:117] "RemoveContainer" containerID="d9c752137312be22ad9cd96fc9b25ccb983a1b5666ce8020a1265aceeb30de90" Nov 29 00:33:07 crc kubenswrapper[4931]: I1129 00:33:07.016633 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4nv4g"] Nov 29 00:33:07 crc kubenswrapper[4931]: I1129 00:33:07.025614 4931 scope.go:117] "RemoveContainer" containerID="64d68d2e05275ca0781abd42f038c6ac48dfcff06b16d6fc0dea82168a63676d" Nov 29 00:33:07 crc kubenswrapper[4931]: I1129 00:33:07.027565 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4nv4g"] Nov 29 00:33:07 crc kubenswrapper[4931]: I1129 00:33:07.064293 4931 scope.go:117] "RemoveContainer" containerID="109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da" Nov 29 00:33:07 crc kubenswrapper[4931]: E1129 00:33:07.064710 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da\": container with ID starting with 109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da not found: ID does not exist" containerID="109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da" Nov 29 00:33:07 crc kubenswrapper[4931]: I1129 00:33:07.064758 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da"} err="failed to get container status \"109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da\": rpc error: code = NotFound desc = could not find container \"109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da\": container with ID starting with 109c3b7f5f519ad540a921639bb5a9f7c5e5f21bf84e660fac7b22490674e9da not found: ID does not exist" Nov 29 00:33:07 crc kubenswrapper[4931]: I1129 00:33:07.064789 4931 scope.go:117] "RemoveContainer" containerID="d9c752137312be22ad9cd96fc9b25ccb983a1b5666ce8020a1265aceeb30de90" Nov 29 00:33:07 crc kubenswrapper[4931]: E1129 00:33:07.065131 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9c752137312be22ad9cd96fc9b25ccb983a1b5666ce8020a1265aceeb30de90\": container with ID starting with d9c752137312be22ad9cd96fc9b25ccb983a1b5666ce8020a1265aceeb30de90 not found: ID does not exist" containerID="d9c752137312be22ad9cd96fc9b25ccb983a1b5666ce8020a1265aceeb30de90" Nov 29 00:33:07 crc kubenswrapper[4931]: I1129 00:33:07.065176 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9c752137312be22ad9cd96fc9b25ccb983a1b5666ce8020a1265aceeb30de90"} err="failed to get container status \"d9c752137312be22ad9cd96fc9b25ccb983a1b5666ce8020a1265aceeb30de90\": rpc error: code = NotFound desc = could not find container \"d9c752137312be22ad9cd96fc9b25ccb983a1b5666ce8020a1265aceeb30de90\": container with ID starting with d9c752137312be22ad9cd96fc9b25ccb983a1b5666ce8020a1265aceeb30de90 not found: ID does not exist" Nov 29 00:33:07 crc kubenswrapper[4931]: I1129 00:33:07.065198 4931 scope.go:117] "RemoveContainer" containerID="64d68d2e05275ca0781abd42f038c6ac48dfcff06b16d6fc0dea82168a63676d" Nov 29 00:33:07 crc kubenswrapper[4931]: E1129 00:33:07.065563 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64d68d2e05275ca0781abd42f038c6ac48dfcff06b16d6fc0dea82168a63676d\": container with ID starting with 64d68d2e05275ca0781abd42f038c6ac48dfcff06b16d6fc0dea82168a63676d not found: ID does not exist" containerID="64d68d2e05275ca0781abd42f038c6ac48dfcff06b16d6fc0dea82168a63676d" Nov 29 00:33:07 crc kubenswrapper[4931]: I1129 00:33:07.065627 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64d68d2e05275ca0781abd42f038c6ac48dfcff06b16d6fc0dea82168a63676d"} err="failed to get container status \"64d68d2e05275ca0781abd42f038c6ac48dfcff06b16d6fc0dea82168a63676d\": rpc error: code = NotFound desc = could not find container \"64d68d2e05275ca0781abd42f038c6ac48dfcff06b16d6fc0dea82168a63676d\": container with ID starting with 64d68d2e05275ca0781abd42f038c6ac48dfcff06b16d6fc0dea82168a63676d not found: ID does not exist" Nov 29 00:33:07 crc kubenswrapper[4931]: I1129 00:33:07.232422 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" path="/var/lib/kubelet/pods/1c47b36f-2cf1-44bb-91af-c16c42a84d22/volumes" Nov 29 00:33:08 crc kubenswrapper[4931]: I1129 00:33:08.993570 4931 generic.go:334] "Generic (PLEG): container finished" podID="935c6e60-45b7-4859-bbd0-b147fa600465" containerID="f6947708df7249911bdbc91496c01f462f0a7e414bf1eefa9e4a043595bedfd1" exitCode=0 Nov 29 00:33:08 crc kubenswrapper[4931]: I1129 00:33:08.993681 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" event={"ID":"935c6e60-45b7-4859-bbd0-b147fa600465","Type":"ContainerDied","Data":"f6947708df7249911bdbc91496c01f462f0a7e414bf1eefa9e4a043595bedfd1"} Nov 29 00:33:10 crc kubenswrapper[4931]: I1129 00:33:10.538613 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:33:10 crc kubenswrapper[4931]: I1129 00:33:10.623053 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw2xx\" (UniqueName: \"kubernetes.io/projected/935c6e60-45b7-4859-bbd0-b147fa600465-kube-api-access-zw2xx\") pod \"935c6e60-45b7-4859-bbd0-b147fa600465\" (UID: \"935c6e60-45b7-4859-bbd0-b147fa600465\") " Nov 29 00:33:10 crc kubenswrapper[4931]: I1129 00:33:10.623428 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/935c6e60-45b7-4859-bbd0-b147fa600465-ssh-key\") pod \"935c6e60-45b7-4859-bbd0-b147fa600465\" (UID: \"935c6e60-45b7-4859-bbd0-b147fa600465\") " Nov 29 00:33:10 crc kubenswrapper[4931]: I1129 00:33:10.623472 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/935c6e60-45b7-4859-bbd0-b147fa600465-inventory\") pod \"935c6e60-45b7-4859-bbd0-b147fa600465\" (UID: \"935c6e60-45b7-4859-bbd0-b147fa600465\") " Nov 29 00:33:10 crc kubenswrapper[4931]: I1129 00:33:10.632638 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/935c6e60-45b7-4859-bbd0-b147fa600465-kube-api-access-zw2xx" (OuterVolumeSpecName: "kube-api-access-zw2xx") pod "935c6e60-45b7-4859-bbd0-b147fa600465" (UID: "935c6e60-45b7-4859-bbd0-b147fa600465"). InnerVolumeSpecName "kube-api-access-zw2xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:33:10 crc kubenswrapper[4931]: I1129 00:33:10.663937 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/935c6e60-45b7-4859-bbd0-b147fa600465-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "935c6e60-45b7-4859-bbd0-b147fa600465" (UID: "935c6e60-45b7-4859-bbd0-b147fa600465"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:33:10 crc kubenswrapper[4931]: I1129 00:33:10.666052 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/935c6e60-45b7-4859-bbd0-b147fa600465-inventory" (OuterVolumeSpecName: "inventory") pod "935c6e60-45b7-4859-bbd0-b147fa600465" (UID: "935c6e60-45b7-4859-bbd0-b147fa600465"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:33:10 crc kubenswrapper[4931]: I1129 00:33:10.727468 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw2xx\" (UniqueName: \"kubernetes.io/projected/935c6e60-45b7-4859-bbd0-b147fa600465-kube-api-access-zw2xx\") on node \"crc\" DevicePath \"\"" Nov 29 00:33:10 crc kubenswrapper[4931]: I1129 00:33:10.727526 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/935c6e60-45b7-4859-bbd0-b147fa600465-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:33:10 crc kubenswrapper[4931]: I1129 00:33:10.727553 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/935c6e60-45b7-4859-bbd0-b147fa600465-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.013020 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" event={"ID":"935c6e60-45b7-4859-bbd0-b147fa600465","Type":"ContainerDied","Data":"6d49edcacc893676280d62d1875fc1382953b17867f916663b08a0f3ae8eba82"} Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.013375 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d49edcacc893676280d62d1875fc1382953b17867f916663b08a0f3ae8eba82" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.013065 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.126235 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp"] Nov 29 00:33:11 crc kubenswrapper[4931]: E1129 00:33:11.126917 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" containerName="extract-content" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.126950 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" containerName="extract-content" Nov 29 00:33:11 crc kubenswrapper[4931]: E1129 00:33:11.126980 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="935c6e60-45b7-4859-bbd0-b147fa600465" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.126994 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="935c6e60-45b7-4859-bbd0-b147fa600465" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 29 00:33:11 crc kubenswrapper[4931]: E1129 00:33:11.127037 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" containerName="registry-server" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.127051 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" containerName="registry-server" Nov 29 00:33:11 crc kubenswrapper[4931]: E1129 00:33:11.127092 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" containerName="extract-utilities" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.127105 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" containerName="extract-utilities" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.127496 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c47b36f-2cf1-44bb-91af-c16c42a84d22" containerName="registry-server" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.127543 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="935c6e60-45b7-4859-bbd0-b147fa600465" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.128584 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.131483 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.131884 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.131912 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.131928 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.154616 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp"] Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.238704 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf7dz\" (UniqueName: \"kubernetes.io/projected/3d56a6b7-f516-4675-b2bc-50372feb503d-kube-api-access-gf7dz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp\" (UID: \"3d56a6b7-f516-4675-b2bc-50372feb503d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.238977 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d56a6b7-f516-4675-b2bc-50372feb503d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp\" (UID: \"3d56a6b7-f516-4675-b2bc-50372feb503d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.239242 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d56a6b7-f516-4675-b2bc-50372feb503d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp\" (UID: \"3d56a6b7-f516-4675-b2bc-50372feb503d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.341472 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf7dz\" (UniqueName: \"kubernetes.io/projected/3d56a6b7-f516-4675-b2bc-50372feb503d-kube-api-access-gf7dz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp\" (UID: \"3d56a6b7-f516-4675-b2bc-50372feb503d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.341551 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d56a6b7-f516-4675-b2bc-50372feb503d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp\" (UID: \"3d56a6b7-f516-4675-b2bc-50372feb503d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.343094 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d56a6b7-f516-4675-b2bc-50372feb503d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp\" (UID: \"3d56a6b7-f516-4675-b2bc-50372feb503d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.348531 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d56a6b7-f516-4675-b2bc-50372feb503d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp\" (UID: \"3d56a6b7-f516-4675-b2bc-50372feb503d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.350086 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d56a6b7-f516-4675-b2bc-50372feb503d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp\" (UID: \"3d56a6b7-f516-4675-b2bc-50372feb503d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.373923 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf7dz\" (UniqueName: \"kubernetes.io/projected/3d56a6b7-f516-4675-b2bc-50372feb503d-kube-api-access-gf7dz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp\" (UID: \"3d56a6b7-f516-4675-b2bc-50372feb503d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:33:11 crc kubenswrapper[4931]: I1129 00:33:11.452682 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:33:12 crc kubenswrapper[4931]: I1129 00:33:12.039517 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp"] Nov 29 00:33:13 crc kubenswrapper[4931]: I1129 00:33:13.033207 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" event={"ID":"3d56a6b7-f516-4675-b2bc-50372feb503d","Type":"ContainerStarted","Data":"1b76e5d4e3ba3537cee8dbdefa3a0cd9ac42b7fe809d3e11d06a3942685c2f95"} Nov 29 00:33:13 crc kubenswrapper[4931]: I1129 00:33:13.033733 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" event={"ID":"3d56a6b7-f516-4675-b2bc-50372feb503d","Type":"ContainerStarted","Data":"c5f578776ec5bf371732c38f75382921839fc8e37ab8e46d3ac9771a5ea89810"} Nov 29 00:33:13 crc kubenswrapper[4931]: I1129 00:33:13.068159 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" podStartSLOduration=1.6001715239999998 podStartE2EDuration="2.068137288s" podCreationTimestamp="2025-11-29 00:33:11 +0000 UTC" firstStartedPulling="2025-11-29 00:33:12.053867225 +0000 UTC m=+2005.215760457" lastFinishedPulling="2025-11-29 00:33:12.521832989 +0000 UTC m=+2005.683726221" observedRunningTime="2025-11-29 00:33:13.054356412 +0000 UTC m=+2006.216249694" watchObservedRunningTime="2025-11-29 00:33:13.068137288 +0000 UTC m=+2006.230030530" Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.384713 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n7fsh"] Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.388175 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.408084 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n7fsh"] Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.460802 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/468f956f-fb03-405b-ade7-46919e93281e-catalog-content\") pod \"certified-operators-n7fsh\" (UID: \"468f956f-fb03-405b-ade7-46919e93281e\") " pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.460916 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/468f956f-fb03-405b-ade7-46919e93281e-utilities\") pod \"certified-operators-n7fsh\" (UID: \"468f956f-fb03-405b-ade7-46919e93281e\") " pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.460975 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c568j\" (UniqueName: \"kubernetes.io/projected/468f956f-fb03-405b-ade7-46919e93281e-kube-api-access-c568j\") pod \"certified-operators-n7fsh\" (UID: \"468f956f-fb03-405b-ade7-46919e93281e\") " pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.563317 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/468f956f-fb03-405b-ade7-46919e93281e-catalog-content\") pod \"certified-operators-n7fsh\" (UID: \"468f956f-fb03-405b-ade7-46919e93281e\") " pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.563710 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/468f956f-fb03-405b-ade7-46919e93281e-utilities\") pod \"certified-operators-n7fsh\" (UID: \"468f956f-fb03-405b-ade7-46919e93281e\") " pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.563773 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c568j\" (UniqueName: \"kubernetes.io/projected/468f956f-fb03-405b-ade7-46919e93281e-kube-api-access-c568j\") pod \"certified-operators-n7fsh\" (UID: \"468f956f-fb03-405b-ade7-46919e93281e\") " pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.564248 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/468f956f-fb03-405b-ade7-46919e93281e-catalog-content\") pod \"certified-operators-n7fsh\" (UID: \"468f956f-fb03-405b-ade7-46919e93281e\") " pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.564445 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/468f956f-fb03-405b-ade7-46919e93281e-utilities\") pod \"certified-operators-n7fsh\" (UID: \"468f956f-fb03-405b-ade7-46919e93281e\") " pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.590683 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c568j\" (UniqueName: \"kubernetes.io/projected/468f956f-fb03-405b-ade7-46919e93281e-kube-api-access-c568j\") pod \"certified-operators-n7fsh\" (UID: \"468f956f-fb03-405b-ade7-46919e93281e\") " pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:01 crc kubenswrapper[4931]: I1129 00:34:01.724828 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:02 crc kubenswrapper[4931]: W1129 00:34:02.292154 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod468f956f_fb03_405b_ade7_46919e93281e.slice/crio-385361e65c61b28c629117bd6a90c770c8c852881149bde979420083de4d23c1 WatchSource:0}: Error finding container 385361e65c61b28c629117bd6a90c770c8c852881149bde979420083de4d23c1: Status 404 returned error can't find the container with id 385361e65c61b28c629117bd6a90c770c8c852881149bde979420083de4d23c1 Nov 29 00:34:02 crc kubenswrapper[4931]: I1129 00:34:02.296213 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n7fsh"] Nov 29 00:34:02 crc kubenswrapper[4931]: I1129 00:34:02.565367 4931 generic.go:334] "Generic (PLEG): container finished" podID="468f956f-fb03-405b-ade7-46919e93281e" containerID="c63129adc265cddfe084643a7890b200411511049db208402c9f9958e322e5bf" exitCode=0 Nov 29 00:34:02 crc kubenswrapper[4931]: I1129 00:34:02.565662 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7fsh" event={"ID":"468f956f-fb03-405b-ade7-46919e93281e","Type":"ContainerDied","Data":"c63129adc265cddfe084643a7890b200411511049db208402c9f9958e322e5bf"} Nov 29 00:34:02 crc kubenswrapper[4931]: I1129 00:34:02.565691 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7fsh" event={"ID":"468f956f-fb03-405b-ade7-46919e93281e","Type":"ContainerStarted","Data":"385361e65c61b28c629117bd6a90c770c8c852881149bde979420083de4d23c1"} Nov 29 00:34:03 crc kubenswrapper[4931]: I1129 00:34:03.581495 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7fsh" event={"ID":"468f956f-fb03-405b-ade7-46919e93281e","Type":"ContainerStarted","Data":"32a3072cb42d291c32130cf2af084301795875b12d28133e5a63e517c62710bb"} Nov 29 00:34:04 crc kubenswrapper[4931]: I1129 00:34:04.593510 4931 generic.go:334] "Generic (PLEG): container finished" podID="468f956f-fb03-405b-ade7-46919e93281e" containerID="32a3072cb42d291c32130cf2af084301795875b12d28133e5a63e517c62710bb" exitCode=0 Nov 29 00:34:04 crc kubenswrapper[4931]: I1129 00:34:04.593566 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7fsh" event={"ID":"468f956f-fb03-405b-ade7-46919e93281e","Type":"ContainerDied","Data":"32a3072cb42d291c32130cf2af084301795875b12d28133e5a63e517c62710bb"} Nov 29 00:34:05 crc kubenswrapper[4931]: I1129 00:34:05.604616 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7fsh" event={"ID":"468f956f-fb03-405b-ade7-46919e93281e","Type":"ContainerStarted","Data":"18c9275c0835d2490eda45591ab92fdbf9451e3244532cd799177a505d3db268"} Nov 29 00:34:05 crc kubenswrapper[4931]: I1129 00:34:05.628910 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n7fsh" podStartSLOduration=2.12260359 podStartE2EDuration="4.628889823s" podCreationTimestamp="2025-11-29 00:34:01 +0000 UTC" firstStartedPulling="2025-11-29 00:34:02.567440001 +0000 UTC m=+2055.729333243" lastFinishedPulling="2025-11-29 00:34:05.073726234 +0000 UTC m=+2058.235619476" observedRunningTime="2025-11-29 00:34:05.62271534 +0000 UTC m=+2058.784608582" watchObservedRunningTime="2025-11-29 00:34:05.628889823 +0000 UTC m=+2058.790783075" Nov 29 00:34:11 crc kubenswrapper[4931]: I1129 00:34:11.726246 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:11 crc kubenswrapper[4931]: I1129 00:34:11.726776 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:11 crc kubenswrapper[4931]: I1129 00:34:11.794321 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:12 crc kubenswrapper[4931]: I1129 00:34:12.757136 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:12 crc kubenswrapper[4931]: I1129 00:34:12.824851 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n7fsh"] Nov 29 00:34:14 crc kubenswrapper[4931]: I1129 00:34:14.697977 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n7fsh" podUID="468f956f-fb03-405b-ade7-46919e93281e" containerName="registry-server" containerID="cri-o://18c9275c0835d2490eda45591ab92fdbf9451e3244532cd799177a505d3db268" gracePeriod=2 Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.710956 4931 generic.go:334] "Generic (PLEG): container finished" podID="468f956f-fb03-405b-ade7-46919e93281e" containerID="18c9275c0835d2490eda45591ab92fdbf9451e3244532cd799177a505d3db268" exitCode=0 Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.711009 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7fsh" event={"ID":"468f956f-fb03-405b-ade7-46919e93281e","Type":"ContainerDied","Data":"18c9275c0835d2490eda45591ab92fdbf9451e3244532cd799177a505d3db268"} Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.711542 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7fsh" event={"ID":"468f956f-fb03-405b-ade7-46919e93281e","Type":"ContainerDied","Data":"385361e65c61b28c629117bd6a90c770c8c852881149bde979420083de4d23c1"} Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.711561 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="385361e65c61b28c629117bd6a90c770c8c852881149bde979420083de4d23c1" Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.737610 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.792773 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/468f956f-fb03-405b-ade7-46919e93281e-catalog-content\") pod \"468f956f-fb03-405b-ade7-46919e93281e\" (UID: \"468f956f-fb03-405b-ade7-46919e93281e\") " Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.792908 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c568j\" (UniqueName: \"kubernetes.io/projected/468f956f-fb03-405b-ade7-46919e93281e-kube-api-access-c568j\") pod \"468f956f-fb03-405b-ade7-46919e93281e\" (UID: \"468f956f-fb03-405b-ade7-46919e93281e\") " Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.793021 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/468f956f-fb03-405b-ade7-46919e93281e-utilities\") pod \"468f956f-fb03-405b-ade7-46919e93281e\" (UID: \"468f956f-fb03-405b-ade7-46919e93281e\") " Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.794531 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/468f956f-fb03-405b-ade7-46919e93281e-utilities" (OuterVolumeSpecName: "utilities") pod "468f956f-fb03-405b-ade7-46919e93281e" (UID: "468f956f-fb03-405b-ade7-46919e93281e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.802489 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/468f956f-fb03-405b-ade7-46919e93281e-kube-api-access-c568j" (OuterVolumeSpecName: "kube-api-access-c568j") pod "468f956f-fb03-405b-ade7-46919e93281e" (UID: "468f956f-fb03-405b-ade7-46919e93281e"). InnerVolumeSpecName "kube-api-access-c568j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.854169 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/468f956f-fb03-405b-ade7-46919e93281e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "468f956f-fb03-405b-ade7-46919e93281e" (UID: "468f956f-fb03-405b-ade7-46919e93281e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.895119 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/468f956f-fb03-405b-ade7-46919e93281e-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.895152 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/468f956f-fb03-405b-ade7-46919e93281e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:34:15 crc kubenswrapper[4931]: I1129 00:34:15.895163 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c568j\" (UniqueName: \"kubernetes.io/projected/468f956f-fb03-405b-ade7-46919e93281e-kube-api-access-c568j\") on node \"crc\" DevicePath \"\"" Nov 29 00:34:16 crc kubenswrapper[4931]: I1129 00:34:16.719120 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n7fsh" Nov 29 00:34:16 crc kubenswrapper[4931]: I1129 00:34:16.752309 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n7fsh"] Nov 29 00:34:16 crc kubenswrapper[4931]: I1129 00:34:16.759058 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n7fsh"] Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.227299 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="468f956f-fb03-405b-ade7-46919e93281e" path="/var/lib/kubelet/pods/468f956f-fb03-405b-ade7-46919e93281e/volumes" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.652523 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p6wrw"] Nov 29 00:34:17 crc kubenswrapper[4931]: E1129 00:34:17.653272 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="468f956f-fb03-405b-ade7-46919e93281e" containerName="extract-utilities" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.653290 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="468f956f-fb03-405b-ade7-46919e93281e" containerName="extract-utilities" Nov 29 00:34:17 crc kubenswrapper[4931]: E1129 00:34:17.653306 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="468f956f-fb03-405b-ade7-46919e93281e" containerName="registry-server" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.653312 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="468f956f-fb03-405b-ade7-46919e93281e" containerName="registry-server" Nov 29 00:34:17 crc kubenswrapper[4931]: E1129 00:34:17.653328 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="468f956f-fb03-405b-ade7-46919e93281e" containerName="extract-content" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.653334 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="468f956f-fb03-405b-ade7-46919e93281e" containerName="extract-content" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.653519 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="468f956f-fb03-405b-ade7-46919e93281e" containerName="registry-server" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.654848 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.681846 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p6wrw"] Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.732890 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj6rc\" (UniqueName: \"kubernetes.io/projected/aedec608-c49a-4afb-879e-5385481ea961-kube-api-access-bj6rc\") pod \"redhat-marketplace-p6wrw\" (UID: \"aedec608-c49a-4afb-879e-5385481ea961\") " pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.732935 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aedec608-c49a-4afb-879e-5385481ea961-catalog-content\") pod \"redhat-marketplace-p6wrw\" (UID: \"aedec608-c49a-4afb-879e-5385481ea961\") " pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.732975 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aedec608-c49a-4afb-879e-5385481ea961-utilities\") pod \"redhat-marketplace-p6wrw\" (UID: \"aedec608-c49a-4afb-879e-5385481ea961\") " pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.835414 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aedec608-c49a-4afb-879e-5385481ea961-utilities\") pod \"redhat-marketplace-p6wrw\" (UID: \"aedec608-c49a-4afb-879e-5385481ea961\") " pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.835585 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj6rc\" (UniqueName: \"kubernetes.io/projected/aedec608-c49a-4afb-879e-5385481ea961-kube-api-access-bj6rc\") pod \"redhat-marketplace-p6wrw\" (UID: \"aedec608-c49a-4afb-879e-5385481ea961\") " pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.835613 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aedec608-c49a-4afb-879e-5385481ea961-catalog-content\") pod \"redhat-marketplace-p6wrw\" (UID: \"aedec608-c49a-4afb-879e-5385481ea961\") " pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.836068 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aedec608-c49a-4afb-879e-5385481ea961-catalog-content\") pod \"redhat-marketplace-p6wrw\" (UID: \"aedec608-c49a-4afb-879e-5385481ea961\") " pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.836152 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aedec608-c49a-4afb-879e-5385481ea961-utilities\") pod \"redhat-marketplace-p6wrw\" (UID: \"aedec608-c49a-4afb-879e-5385481ea961\") " pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:17 crc kubenswrapper[4931]: I1129 00:34:17.858751 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj6rc\" (UniqueName: \"kubernetes.io/projected/aedec608-c49a-4afb-879e-5385481ea961-kube-api-access-bj6rc\") pod \"redhat-marketplace-p6wrw\" (UID: \"aedec608-c49a-4afb-879e-5385481ea961\") " pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:18 crc kubenswrapper[4931]: I1129 00:34:18.009673 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:18 crc kubenswrapper[4931]: I1129 00:34:18.437994 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p6wrw"] Nov 29 00:34:18 crc kubenswrapper[4931]: I1129 00:34:18.737407 4931 generic.go:334] "Generic (PLEG): container finished" podID="aedec608-c49a-4afb-879e-5385481ea961" containerID="eb7aebe693b909ffe2e601e4a6b3a93426c89f9070de29a4839eb17128bc67de" exitCode=0 Nov 29 00:34:18 crc kubenswrapper[4931]: I1129 00:34:18.737504 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6wrw" event={"ID":"aedec608-c49a-4afb-879e-5385481ea961","Type":"ContainerDied","Data":"eb7aebe693b909ffe2e601e4a6b3a93426c89f9070de29a4839eb17128bc67de"} Nov 29 00:34:18 crc kubenswrapper[4931]: I1129 00:34:18.737779 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6wrw" event={"ID":"aedec608-c49a-4afb-879e-5385481ea961","Type":"ContainerStarted","Data":"de82312b704faa9499afb811c262c6e7d5cab159bbb16aed685f63d971d06db0"} Nov 29 00:34:20 crc kubenswrapper[4931]: I1129 00:34:20.767110 4931 generic.go:334] "Generic (PLEG): container finished" podID="aedec608-c49a-4afb-879e-5385481ea961" containerID="a118595c710149664167d91ee0759edb213c561c780a716c10c5dc4d56221627" exitCode=0 Nov 29 00:34:20 crc kubenswrapper[4931]: I1129 00:34:20.767199 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6wrw" event={"ID":"aedec608-c49a-4afb-879e-5385481ea961","Type":"ContainerDied","Data":"a118595c710149664167d91ee0759edb213c561c780a716c10c5dc4d56221627"} Nov 29 00:34:21 crc kubenswrapper[4931]: I1129 00:34:21.784847 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6wrw" event={"ID":"aedec608-c49a-4afb-879e-5385481ea961","Type":"ContainerStarted","Data":"39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869"} Nov 29 00:34:21 crc kubenswrapper[4931]: I1129 00:34:21.818997 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p6wrw" podStartSLOduration=2.259705497 podStartE2EDuration="4.818975158s" podCreationTimestamp="2025-11-29 00:34:17 +0000 UTC" firstStartedPulling="2025-11-29 00:34:18.739047606 +0000 UTC m=+2071.900940838" lastFinishedPulling="2025-11-29 00:34:21.298317267 +0000 UTC m=+2074.460210499" observedRunningTime="2025-11-29 00:34:21.808459442 +0000 UTC m=+2074.970352704" watchObservedRunningTime="2025-11-29 00:34:21.818975158 +0000 UTC m=+2074.980868390" Nov 29 00:34:28 crc kubenswrapper[4931]: I1129 00:34:28.010338 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:28 crc kubenswrapper[4931]: I1129 00:34:28.011175 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:28 crc kubenswrapper[4931]: I1129 00:34:28.064998 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:28 crc kubenswrapper[4931]: I1129 00:34:28.908850 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:28 crc kubenswrapper[4931]: I1129 00:34:28.955361 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p6wrw"] Nov 29 00:34:30 crc kubenswrapper[4931]: I1129 00:34:30.882760 4931 generic.go:334] "Generic (PLEG): container finished" podID="3d56a6b7-f516-4675-b2bc-50372feb503d" containerID="1b76e5d4e3ba3537cee8dbdefa3a0cd9ac42b7fe809d3e11d06a3942685c2f95" exitCode=0 Nov 29 00:34:30 crc kubenswrapper[4931]: I1129 00:34:30.882845 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" event={"ID":"3d56a6b7-f516-4675-b2bc-50372feb503d","Type":"ContainerDied","Data":"1b76e5d4e3ba3537cee8dbdefa3a0cd9ac42b7fe809d3e11d06a3942685c2f95"} Nov 29 00:34:30 crc kubenswrapper[4931]: I1129 00:34:30.883107 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p6wrw" podUID="aedec608-c49a-4afb-879e-5385481ea961" containerName="registry-server" containerID="cri-o://39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869" gracePeriod=2 Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.384088 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.518736 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj6rc\" (UniqueName: \"kubernetes.io/projected/aedec608-c49a-4afb-879e-5385481ea961-kube-api-access-bj6rc\") pod \"aedec608-c49a-4afb-879e-5385481ea961\" (UID: \"aedec608-c49a-4afb-879e-5385481ea961\") " Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.518907 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aedec608-c49a-4afb-879e-5385481ea961-utilities\") pod \"aedec608-c49a-4afb-879e-5385481ea961\" (UID: \"aedec608-c49a-4afb-879e-5385481ea961\") " Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.519003 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aedec608-c49a-4afb-879e-5385481ea961-catalog-content\") pod \"aedec608-c49a-4afb-879e-5385481ea961\" (UID: \"aedec608-c49a-4afb-879e-5385481ea961\") " Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.520089 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aedec608-c49a-4afb-879e-5385481ea961-utilities" (OuterVolumeSpecName: "utilities") pod "aedec608-c49a-4afb-879e-5385481ea961" (UID: "aedec608-c49a-4afb-879e-5385481ea961"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.527838 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aedec608-c49a-4afb-879e-5385481ea961-kube-api-access-bj6rc" (OuterVolumeSpecName: "kube-api-access-bj6rc") pod "aedec608-c49a-4afb-879e-5385481ea961" (UID: "aedec608-c49a-4afb-879e-5385481ea961"). InnerVolumeSpecName "kube-api-access-bj6rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.538001 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aedec608-c49a-4afb-879e-5385481ea961-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aedec608-c49a-4afb-879e-5385481ea961" (UID: "aedec608-c49a-4afb-879e-5385481ea961"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.622000 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj6rc\" (UniqueName: \"kubernetes.io/projected/aedec608-c49a-4afb-879e-5385481ea961-kube-api-access-bj6rc\") on node \"crc\" DevicePath \"\"" Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.622281 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aedec608-c49a-4afb-879e-5385481ea961-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.622339 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aedec608-c49a-4afb-879e-5385481ea961-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.899618 4931 generic.go:334] "Generic (PLEG): container finished" podID="aedec608-c49a-4afb-879e-5385481ea961" containerID="39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869" exitCode=0 Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.899676 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p6wrw" Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.899728 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6wrw" event={"ID":"aedec608-c49a-4afb-879e-5385481ea961","Type":"ContainerDied","Data":"39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869"} Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.899761 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p6wrw" event={"ID":"aedec608-c49a-4afb-879e-5385481ea961","Type":"ContainerDied","Data":"de82312b704faa9499afb811c262c6e7d5cab159bbb16aed685f63d971d06db0"} Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.899780 4931 scope.go:117] "RemoveContainer" containerID="39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869" Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.922874 4931 scope.go:117] "RemoveContainer" containerID="a118595c710149664167d91ee0759edb213c561c780a716c10c5dc4d56221627" Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.974829 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p6wrw"] Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.975622 4931 scope.go:117] "RemoveContainer" containerID="eb7aebe693b909ffe2e601e4a6b3a93426c89f9070de29a4839eb17128bc67de" Nov 29 00:34:31 crc kubenswrapper[4931]: I1129 00:34:31.994662 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p6wrw"] Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.008625 4931 scope.go:117] "RemoveContainer" containerID="39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869" Nov 29 00:34:32 crc kubenswrapper[4931]: E1129 00:34:32.009953 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869\": container with ID starting with 39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869 not found: ID does not exist" containerID="39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.010019 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869"} err="failed to get container status \"39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869\": rpc error: code = NotFound desc = could not find container \"39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869\": container with ID starting with 39024a8d6b397b740069742e5887014b89deaf84582c87ba059ad3610f553869 not found: ID does not exist" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.010059 4931 scope.go:117] "RemoveContainer" containerID="a118595c710149664167d91ee0759edb213c561c780a716c10c5dc4d56221627" Nov 29 00:34:32 crc kubenswrapper[4931]: E1129 00:34:32.010447 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a118595c710149664167d91ee0759edb213c561c780a716c10c5dc4d56221627\": container with ID starting with a118595c710149664167d91ee0759edb213c561c780a716c10c5dc4d56221627 not found: ID does not exist" containerID="a118595c710149664167d91ee0759edb213c561c780a716c10c5dc4d56221627" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.010486 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a118595c710149664167d91ee0759edb213c561c780a716c10c5dc4d56221627"} err="failed to get container status \"a118595c710149664167d91ee0759edb213c561c780a716c10c5dc4d56221627\": rpc error: code = NotFound desc = could not find container \"a118595c710149664167d91ee0759edb213c561c780a716c10c5dc4d56221627\": container with ID starting with a118595c710149664167d91ee0759edb213c561c780a716c10c5dc4d56221627 not found: ID does not exist" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.010517 4931 scope.go:117] "RemoveContainer" containerID="eb7aebe693b909ffe2e601e4a6b3a93426c89f9070de29a4839eb17128bc67de" Nov 29 00:34:32 crc kubenswrapper[4931]: E1129 00:34:32.011046 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb7aebe693b909ffe2e601e4a6b3a93426c89f9070de29a4839eb17128bc67de\": container with ID starting with eb7aebe693b909ffe2e601e4a6b3a93426c89f9070de29a4839eb17128bc67de not found: ID does not exist" containerID="eb7aebe693b909ffe2e601e4a6b3a93426c89f9070de29a4839eb17128bc67de" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.011076 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb7aebe693b909ffe2e601e4a6b3a93426c89f9070de29a4839eb17128bc67de"} err="failed to get container status \"eb7aebe693b909ffe2e601e4a6b3a93426c89f9070de29a4839eb17128bc67de\": rpc error: code = NotFound desc = could not find container \"eb7aebe693b909ffe2e601e4a6b3a93426c89f9070de29a4839eb17128bc67de\": container with ID starting with eb7aebe693b909ffe2e601e4a6b3a93426c89f9070de29a4839eb17128bc67de not found: ID does not exist" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.357941 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.436603 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf7dz\" (UniqueName: \"kubernetes.io/projected/3d56a6b7-f516-4675-b2bc-50372feb503d-kube-api-access-gf7dz\") pod \"3d56a6b7-f516-4675-b2bc-50372feb503d\" (UID: \"3d56a6b7-f516-4675-b2bc-50372feb503d\") " Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.436715 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d56a6b7-f516-4675-b2bc-50372feb503d-inventory\") pod \"3d56a6b7-f516-4675-b2bc-50372feb503d\" (UID: \"3d56a6b7-f516-4675-b2bc-50372feb503d\") " Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.436876 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d56a6b7-f516-4675-b2bc-50372feb503d-ssh-key\") pod \"3d56a6b7-f516-4675-b2bc-50372feb503d\" (UID: \"3d56a6b7-f516-4675-b2bc-50372feb503d\") " Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.440892 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d56a6b7-f516-4675-b2bc-50372feb503d-kube-api-access-gf7dz" (OuterVolumeSpecName: "kube-api-access-gf7dz") pod "3d56a6b7-f516-4675-b2bc-50372feb503d" (UID: "3d56a6b7-f516-4675-b2bc-50372feb503d"). InnerVolumeSpecName "kube-api-access-gf7dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.467086 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d56a6b7-f516-4675-b2bc-50372feb503d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3d56a6b7-f516-4675-b2bc-50372feb503d" (UID: "3d56a6b7-f516-4675-b2bc-50372feb503d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.471262 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d56a6b7-f516-4675-b2bc-50372feb503d-inventory" (OuterVolumeSpecName: "inventory") pod "3d56a6b7-f516-4675-b2bc-50372feb503d" (UID: "3d56a6b7-f516-4675-b2bc-50372feb503d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.539006 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf7dz\" (UniqueName: \"kubernetes.io/projected/3d56a6b7-f516-4675-b2bc-50372feb503d-kube-api-access-gf7dz\") on node \"crc\" DevicePath \"\"" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.539055 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d56a6b7-f516-4675-b2bc-50372feb503d-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.539067 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d56a6b7-f516-4675-b2bc-50372feb503d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.922177 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" event={"ID":"3d56a6b7-f516-4675-b2bc-50372feb503d","Type":"ContainerDied","Data":"c5f578776ec5bf371732c38f75382921839fc8e37ab8e46d3ac9771a5ea89810"} Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.922226 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5f578776ec5bf371732c38f75382921839fc8e37ab8e46d3ac9771a5ea89810" Nov 29 00:34:32 crc kubenswrapper[4931]: I1129 00:34:32.922320 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.084875 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs"] Nov 29 00:34:33 crc kubenswrapper[4931]: E1129 00:34:33.085863 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aedec608-c49a-4afb-879e-5385481ea961" containerName="extract-content" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.085883 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="aedec608-c49a-4afb-879e-5385481ea961" containerName="extract-content" Nov 29 00:34:33 crc kubenswrapper[4931]: E1129 00:34:33.085922 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d56a6b7-f516-4675-b2bc-50372feb503d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.085932 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d56a6b7-f516-4675-b2bc-50372feb503d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 29 00:34:33 crc kubenswrapper[4931]: E1129 00:34:33.085999 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aedec608-c49a-4afb-879e-5385481ea961" containerName="extract-utilities" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.086016 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="aedec608-c49a-4afb-879e-5385481ea961" containerName="extract-utilities" Nov 29 00:34:33 crc kubenswrapper[4931]: E1129 00:34:33.086045 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aedec608-c49a-4afb-879e-5385481ea961" containerName="registry-server" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.086055 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="aedec608-c49a-4afb-879e-5385481ea961" containerName="registry-server" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.086494 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d56a6b7-f516-4675-b2bc-50372feb503d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.086541 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="aedec608-c49a-4afb-879e-5385481ea961" containerName="registry-server" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.087653 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.096443 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs"] Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.097854 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.097942 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.098177 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.098201 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.150053 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01b89e61-27d6-407a-97e0-7ced1cf90139-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g26zs\" (UID: \"01b89e61-27d6-407a-97e0-7ced1cf90139\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.150486 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccp94\" (UniqueName: \"kubernetes.io/projected/01b89e61-27d6-407a-97e0-7ced1cf90139-kube-api-access-ccp94\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g26zs\" (UID: \"01b89e61-27d6-407a-97e0-7ced1cf90139\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.150937 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01b89e61-27d6-407a-97e0-7ced1cf90139-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g26zs\" (UID: \"01b89e61-27d6-407a-97e0-7ced1cf90139\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.226600 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aedec608-c49a-4afb-879e-5385481ea961" path="/var/lib/kubelet/pods/aedec608-c49a-4afb-879e-5385481ea961/volumes" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.253196 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01b89e61-27d6-407a-97e0-7ced1cf90139-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g26zs\" (UID: \"01b89e61-27d6-407a-97e0-7ced1cf90139\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.253479 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01b89e61-27d6-407a-97e0-7ced1cf90139-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g26zs\" (UID: \"01b89e61-27d6-407a-97e0-7ced1cf90139\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.253743 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccp94\" (UniqueName: \"kubernetes.io/projected/01b89e61-27d6-407a-97e0-7ced1cf90139-kube-api-access-ccp94\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g26zs\" (UID: \"01b89e61-27d6-407a-97e0-7ced1cf90139\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.264432 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01b89e61-27d6-407a-97e0-7ced1cf90139-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g26zs\" (UID: \"01b89e61-27d6-407a-97e0-7ced1cf90139\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.264868 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01b89e61-27d6-407a-97e0-7ced1cf90139-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g26zs\" (UID: \"01b89e61-27d6-407a-97e0-7ced1cf90139\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.273324 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccp94\" (UniqueName: \"kubernetes.io/projected/01b89e61-27d6-407a-97e0-7ced1cf90139-kube-api-access-ccp94\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-g26zs\" (UID: \"01b89e61-27d6-407a-97e0-7ced1cf90139\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.422309 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:33 crc kubenswrapper[4931]: I1129 00:34:33.959982 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs"] Nov 29 00:34:33 crc kubenswrapper[4931]: W1129 00:34:33.965219 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01b89e61_27d6_407a_97e0_7ced1cf90139.slice/crio-25db73ba154d969b65cc242c94aa1351f09a4f4ed4683d1b0260fa012e4ed030 WatchSource:0}: Error finding container 25db73ba154d969b65cc242c94aa1351f09a4f4ed4683d1b0260fa012e4ed030: Status 404 returned error can't find the container with id 25db73ba154d969b65cc242c94aa1351f09a4f4ed4683d1b0260fa012e4ed030 Nov 29 00:34:34 crc kubenswrapper[4931]: I1129 00:34:34.940741 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" event={"ID":"01b89e61-27d6-407a-97e0-7ced1cf90139","Type":"ContainerStarted","Data":"25db73ba154d969b65cc242c94aa1351f09a4f4ed4683d1b0260fa012e4ed030"} Nov 29 00:34:35 crc kubenswrapper[4931]: I1129 00:34:35.952028 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" event={"ID":"01b89e61-27d6-407a-97e0-7ced1cf90139","Type":"ContainerStarted","Data":"3ccf879f411441a553e5d234b3495a1bc846f9fff0762a2bd06e085e913d55bf"} Nov 29 00:34:35 crc kubenswrapper[4931]: I1129 00:34:35.973616 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" podStartSLOduration=2.276767852 podStartE2EDuration="2.973556808s" podCreationTimestamp="2025-11-29 00:34:33 +0000 UTC" firstStartedPulling="2025-11-29 00:34:33.967897816 +0000 UTC m=+2087.129791048" lastFinishedPulling="2025-11-29 00:34:34.664686762 +0000 UTC m=+2087.826580004" observedRunningTime="2025-11-29 00:34:35.965634752 +0000 UTC m=+2089.127527984" watchObservedRunningTime="2025-11-29 00:34:35.973556808 +0000 UTC m=+2089.135450040" Nov 29 00:34:39 crc kubenswrapper[4931]: I1129 00:34:39.992630 4931 generic.go:334] "Generic (PLEG): container finished" podID="01b89e61-27d6-407a-97e0-7ced1cf90139" containerID="3ccf879f411441a553e5d234b3495a1bc846f9fff0762a2bd06e085e913d55bf" exitCode=0 Nov 29 00:34:39 crc kubenswrapper[4931]: I1129 00:34:39.992772 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" event={"ID":"01b89e61-27d6-407a-97e0-7ced1cf90139","Type":"ContainerDied","Data":"3ccf879f411441a553e5d234b3495a1bc846f9fff0762a2bd06e085e913d55bf"} Nov 29 00:34:41 crc kubenswrapper[4931]: I1129 00:34:41.438683 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:41 crc kubenswrapper[4931]: I1129 00:34:41.512725 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccp94\" (UniqueName: \"kubernetes.io/projected/01b89e61-27d6-407a-97e0-7ced1cf90139-kube-api-access-ccp94\") pod \"01b89e61-27d6-407a-97e0-7ced1cf90139\" (UID: \"01b89e61-27d6-407a-97e0-7ced1cf90139\") " Nov 29 00:34:41 crc kubenswrapper[4931]: I1129 00:34:41.512947 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01b89e61-27d6-407a-97e0-7ced1cf90139-ssh-key\") pod \"01b89e61-27d6-407a-97e0-7ced1cf90139\" (UID: \"01b89e61-27d6-407a-97e0-7ced1cf90139\") " Nov 29 00:34:41 crc kubenswrapper[4931]: I1129 00:34:41.513022 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01b89e61-27d6-407a-97e0-7ced1cf90139-inventory\") pod \"01b89e61-27d6-407a-97e0-7ced1cf90139\" (UID: \"01b89e61-27d6-407a-97e0-7ced1cf90139\") " Nov 29 00:34:41 crc kubenswrapper[4931]: I1129 00:34:41.521146 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01b89e61-27d6-407a-97e0-7ced1cf90139-kube-api-access-ccp94" (OuterVolumeSpecName: "kube-api-access-ccp94") pod "01b89e61-27d6-407a-97e0-7ced1cf90139" (UID: "01b89e61-27d6-407a-97e0-7ced1cf90139"). InnerVolumeSpecName "kube-api-access-ccp94". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:34:41 crc kubenswrapper[4931]: I1129 00:34:41.549257 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01b89e61-27d6-407a-97e0-7ced1cf90139-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "01b89e61-27d6-407a-97e0-7ced1cf90139" (UID: "01b89e61-27d6-407a-97e0-7ced1cf90139"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:34:41 crc kubenswrapper[4931]: I1129 00:34:41.553370 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01b89e61-27d6-407a-97e0-7ced1cf90139-inventory" (OuterVolumeSpecName: "inventory") pod "01b89e61-27d6-407a-97e0-7ced1cf90139" (UID: "01b89e61-27d6-407a-97e0-7ced1cf90139"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:34:41 crc kubenswrapper[4931]: I1129 00:34:41.616137 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01b89e61-27d6-407a-97e0-7ced1cf90139-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:34:41 crc kubenswrapper[4931]: I1129 00:34:41.616479 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccp94\" (UniqueName: \"kubernetes.io/projected/01b89e61-27d6-407a-97e0-7ced1cf90139-kube-api-access-ccp94\") on node \"crc\" DevicePath \"\"" Nov 29 00:34:41 crc kubenswrapper[4931]: I1129 00:34:41.616499 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01b89e61-27d6-407a-97e0-7ced1cf90139-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.017038 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" event={"ID":"01b89e61-27d6-407a-97e0-7ced1cf90139","Type":"ContainerDied","Data":"25db73ba154d969b65cc242c94aa1351f09a4f4ed4683d1b0260fa012e4ed030"} Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.017085 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25db73ba154d969b65cc242c94aa1351f09a4f4ed4683d1b0260fa012e4ed030" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.017085 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-g26zs" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.108234 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j"] Nov 29 00:34:42 crc kubenswrapper[4931]: E1129 00:34:42.108757 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01b89e61-27d6-407a-97e0-7ced1cf90139" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.108781 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="01b89e61-27d6-407a-97e0-7ced1cf90139" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.109059 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="01b89e61-27d6-407a-97e0-7ced1cf90139" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.109887 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.113694 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.113848 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.113877 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.114076 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.118850 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j"] Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.230139 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb09a671-9026-4aec-ba97-911979c3f0d7-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kgz5j\" (UID: \"cb09a671-9026-4aec-ba97-911979c3f0d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.230200 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zshhp\" (UniqueName: \"kubernetes.io/projected/cb09a671-9026-4aec-ba97-911979c3f0d7-kube-api-access-zshhp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kgz5j\" (UID: \"cb09a671-9026-4aec-ba97-911979c3f0d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.230242 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb09a671-9026-4aec-ba97-911979c3f0d7-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kgz5j\" (UID: \"cb09a671-9026-4aec-ba97-911979c3f0d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.332242 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb09a671-9026-4aec-ba97-911979c3f0d7-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kgz5j\" (UID: \"cb09a671-9026-4aec-ba97-911979c3f0d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.332312 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zshhp\" (UniqueName: \"kubernetes.io/projected/cb09a671-9026-4aec-ba97-911979c3f0d7-kube-api-access-zshhp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kgz5j\" (UID: \"cb09a671-9026-4aec-ba97-911979c3f0d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.332363 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb09a671-9026-4aec-ba97-911979c3f0d7-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kgz5j\" (UID: \"cb09a671-9026-4aec-ba97-911979c3f0d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.336144 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb09a671-9026-4aec-ba97-911979c3f0d7-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kgz5j\" (UID: \"cb09a671-9026-4aec-ba97-911979c3f0d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.337575 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb09a671-9026-4aec-ba97-911979c3f0d7-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kgz5j\" (UID: \"cb09a671-9026-4aec-ba97-911979c3f0d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.361839 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zshhp\" (UniqueName: \"kubernetes.io/projected/cb09a671-9026-4aec-ba97-911979c3f0d7-kube-api-access-zshhp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kgz5j\" (UID: \"cb09a671-9026-4aec-ba97-911979c3f0d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.434173 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:34:42 crc kubenswrapper[4931]: I1129 00:34:42.961170 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j"] Nov 29 00:34:43 crc kubenswrapper[4931]: I1129 00:34:43.028101 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" event={"ID":"cb09a671-9026-4aec-ba97-911979c3f0d7","Type":"ContainerStarted","Data":"3d7c26ce2f3ee1a97f900efc690afa827dd564011b4e2b27119247e8d459aa43"} Nov 29 00:34:44 crc kubenswrapper[4931]: I1129 00:34:44.039869 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" event={"ID":"cb09a671-9026-4aec-ba97-911979c3f0d7","Type":"ContainerStarted","Data":"227eb722ab1af6714486b96cddf3bb2251edee79bd9ccfe54d668647ce766d9d"} Nov 29 00:34:44 crc kubenswrapper[4931]: I1129 00:34:44.070473 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" podStartSLOduration=1.488128653 podStartE2EDuration="2.070439682s" podCreationTimestamp="2025-11-29 00:34:42 +0000 UTC" firstStartedPulling="2025-11-29 00:34:42.974804921 +0000 UTC m=+2096.136698153" lastFinishedPulling="2025-11-29 00:34:43.55711594 +0000 UTC m=+2096.719009182" observedRunningTime="2025-11-29 00:34:44.055562157 +0000 UTC m=+2097.217455489" watchObservedRunningTime="2025-11-29 00:34:44.070439682 +0000 UTC m=+2097.232332954" Nov 29 00:35:23 crc kubenswrapper[4931]: I1129 00:35:23.734881 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:35:23 crc kubenswrapper[4931]: I1129 00:35:23.735404 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:35:25 crc kubenswrapper[4931]: I1129 00:35:25.407312 4931 generic.go:334] "Generic (PLEG): container finished" podID="cb09a671-9026-4aec-ba97-911979c3f0d7" containerID="227eb722ab1af6714486b96cddf3bb2251edee79bd9ccfe54d668647ce766d9d" exitCode=0 Nov 29 00:35:25 crc kubenswrapper[4931]: I1129 00:35:25.407388 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" event={"ID":"cb09a671-9026-4aec-ba97-911979c3f0d7","Type":"ContainerDied","Data":"227eb722ab1af6714486b96cddf3bb2251edee79bd9ccfe54d668647ce766d9d"} Nov 29 00:35:26 crc kubenswrapper[4931]: I1129 00:35:26.988940 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.103165 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb09a671-9026-4aec-ba97-911979c3f0d7-ssh-key\") pod \"cb09a671-9026-4aec-ba97-911979c3f0d7\" (UID: \"cb09a671-9026-4aec-ba97-911979c3f0d7\") " Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.103324 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb09a671-9026-4aec-ba97-911979c3f0d7-inventory\") pod \"cb09a671-9026-4aec-ba97-911979c3f0d7\" (UID: \"cb09a671-9026-4aec-ba97-911979c3f0d7\") " Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.103395 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zshhp\" (UniqueName: \"kubernetes.io/projected/cb09a671-9026-4aec-ba97-911979c3f0d7-kube-api-access-zshhp\") pod \"cb09a671-9026-4aec-ba97-911979c3f0d7\" (UID: \"cb09a671-9026-4aec-ba97-911979c3f0d7\") " Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.108283 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb09a671-9026-4aec-ba97-911979c3f0d7-kube-api-access-zshhp" (OuterVolumeSpecName: "kube-api-access-zshhp") pod "cb09a671-9026-4aec-ba97-911979c3f0d7" (UID: "cb09a671-9026-4aec-ba97-911979c3f0d7"). InnerVolumeSpecName "kube-api-access-zshhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.131750 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb09a671-9026-4aec-ba97-911979c3f0d7-inventory" (OuterVolumeSpecName: "inventory") pod "cb09a671-9026-4aec-ba97-911979c3f0d7" (UID: "cb09a671-9026-4aec-ba97-911979c3f0d7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.133626 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb09a671-9026-4aec-ba97-911979c3f0d7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cb09a671-9026-4aec-ba97-911979c3f0d7" (UID: "cb09a671-9026-4aec-ba97-911979c3f0d7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.206423 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb09a671-9026-4aec-ba97-911979c3f0d7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.206456 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb09a671-9026-4aec-ba97-911979c3f0d7-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.206470 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zshhp\" (UniqueName: \"kubernetes.io/projected/cb09a671-9026-4aec-ba97-911979c3f0d7-kube-api-access-zshhp\") on node \"crc\" DevicePath \"\"" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.433679 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" event={"ID":"cb09a671-9026-4aec-ba97-911979c3f0d7","Type":"ContainerDied","Data":"3d7c26ce2f3ee1a97f900efc690afa827dd564011b4e2b27119247e8d459aa43"} Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.433734 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d7c26ce2f3ee1a97f900efc690afa827dd564011b4e2b27119247e8d459aa43" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.433964 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kgz5j" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.545154 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq"] Nov 29 00:35:27 crc kubenswrapper[4931]: E1129 00:35:27.545574 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb09a671-9026-4aec-ba97-911979c3f0d7" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.545599 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb09a671-9026-4aec-ba97-911979c3f0d7" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.545874 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb09a671-9026-4aec-ba97-911979c3f0d7" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.546680 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.548656 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.548900 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.549344 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.549873 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.563735 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq"] Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.715720 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-brwbq\" (UID: \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.716060 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjrjv\" (UniqueName: \"kubernetes.io/projected/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-kube-api-access-bjrjv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-brwbq\" (UID: \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.716104 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-brwbq\" (UID: \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.818205 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-brwbq\" (UID: \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.818380 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjrjv\" (UniqueName: \"kubernetes.io/projected/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-kube-api-access-bjrjv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-brwbq\" (UID: \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.818434 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-brwbq\" (UID: \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.824043 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-brwbq\" (UID: \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.833049 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-brwbq\" (UID: \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.841270 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjrjv\" (UniqueName: \"kubernetes.io/projected/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-kube-api-access-bjrjv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-brwbq\" (UID: \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:35:27 crc kubenswrapper[4931]: I1129 00:35:27.874397 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:35:28 crc kubenswrapper[4931]: I1129 00:35:28.430886 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq"] Nov 29 00:35:29 crc kubenswrapper[4931]: I1129 00:35:29.459650 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" event={"ID":"5da7ac90-6a60-4473-b8b8-5531d4c2ac68","Type":"ContainerStarted","Data":"a2a604325fc7447492aff40493aa7521f88a5850584df6c4ce79e0309ff3d7b3"} Nov 29 00:35:29 crc kubenswrapper[4931]: I1129 00:35:29.460210 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" event={"ID":"5da7ac90-6a60-4473-b8b8-5531d4c2ac68","Type":"ContainerStarted","Data":"8ef65c183c01bac36c421c67fbe0f286f01a224b692b227269942559241e63e1"} Nov 29 00:35:29 crc kubenswrapper[4931]: I1129 00:35:29.484681 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" podStartSLOduration=2.02500477 podStartE2EDuration="2.484662351s" podCreationTimestamp="2025-11-29 00:35:27 +0000 UTC" firstStartedPulling="2025-11-29 00:35:28.442310999 +0000 UTC m=+2141.604204231" lastFinishedPulling="2025-11-29 00:35:28.90196854 +0000 UTC m=+2142.063861812" observedRunningTime="2025-11-29 00:35:29.480171508 +0000 UTC m=+2142.642064770" watchObservedRunningTime="2025-11-29 00:35:29.484662351 +0000 UTC m=+2142.646555583" Nov 29 00:35:53 crc kubenswrapper[4931]: I1129 00:35:53.734590 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:35:53 crc kubenswrapper[4931]: I1129 00:35:53.736938 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.184062 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b7v2z"] Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.189148 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.202329 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b7v2z"] Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.231260 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g757g\" (UniqueName: \"kubernetes.io/projected/23f6dae0-af26-46d1-851e-9401172634ac-kube-api-access-g757g\") pod \"community-operators-b7v2z\" (UID: \"23f6dae0-af26-46d1-851e-9401172634ac\") " pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.231337 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f6dae0-af26-46d1-851e-9401172634ac-utilities\") pod \"community-operators-b7v2z\" (UID: \"23f6dae0-af26-46d1-851e-9401172634ac\") " pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.231374 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f6dae0-af26-46d1-851e-9401172634ac-catalog-content\") pod \"community-operators-b7v2z\" (UID: \"23f6dae0-af26-46d1-851e-9401172634ac\") " pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.333099 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g757g\" (UniqueName: \"kubernetes.io/projected/23f6dae0-af26-46d1-851e-9401172634ac-kube-api-access-g757g\") pod \"community-operators-b7v2z\" (UID: \"23f6dae0-af26-46d1-851e-9401172634ac\") " pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.333162 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f6dae0-af26-46d1-851e-9401172634ac-utilities\") pod \"community-operators-b7v2z\" (UID: \"23f6dae0-af26-46d1-851e-9401172634ac\") " pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.333195 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f6dae0-af26-46d1-851e-9401172634ac-catalog-content\") pod \"community-operators-b7v2z\" (UID: \"23f6dae0-af26-46d1-851e-9401172634ac\") " pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.333649 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f6dae0-af26-46d1-851e-9401172634ac-catalog-content\") pod \"community-operators-b7v2z\" (UID: \"23f6dae0-af26-46d1-851e-9401172634ac\") " pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.333791 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f6dae0-af26-46d1-851e-9401172634ac-utilities\") pod \"community-operators-b7v2z\" (UID: \"23f6dae0-af26-46d1-851e-9401172634ac\") " pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.360615 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g757g\" (UniqueName: \"kubernetes.io/projected/23f6dae0-af26-46d1-851e-9401172634ac-kube-api-access-g757g\") pod \"community-operators-b7v2z\" (UID: \"23f6dae0-af26-46d1-851e-9401172634ac\") " pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:01 crc kubenswrapper[4931]: I1129 00:36:01.518948 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:02 crc kubenswrapper[4931]: I1129 00:36:02.067226 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b7v2z"] Nov 29 00:36:02 crc kubenswrapper[4931]: I1129 00:36:02.773224 4931 generic.go:334] "Generic (PLEG): container finished" podID="23f6dae0-af26-46d1-851e-9401172634ac" containerID="01d136690bc502ca6ef7b708da2df7f3a468eaddc84aadc4ab46d403e4881f7b" exitCode=0 Nov 29 00:36:02 crc kubenswrapper[4931]: I1129 00:36:02.773356 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b7v2z" event={"ID":"23f6dae0-af26-46d1-851e-9401172634ac","Type":"ContainerDied","Data":"01d136690bc502ca6ef7b708da2df7f3a468eaddc84aadc4ab46d403e4881f7b"} Nov 29 00:36:02 crc kubenswrapper[4931]: I1129 00:36:02.773834 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b7v2z" event={"ID":"23f6dae0-af26-46d1-851e-9401172634ac","Type":"ContainerStarted","Data":"abf16e8e46d7bb6591192783e0615846057f232460969efccfc6a17c4230fb37"} Nov 29 00:36:04 crc kubenswrapper[4931]: I1129 00:36:04.793316 4931 generic.go:334] "Generic (PLEG): container finished" podID="23f6dae0-af26-46d1-851e-9401172634ac" containerID="f400965294df83da970312c01542f42ba08d96b012edb3573480db5c8b7b0480" exitCode=0 Nov 29 00:36:04 crc kubenswrapper[4931]: I1129 00:36:04.793448 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b7v2z" event={"ID":"23f6dae0-af26-46d1-851e-9401172634ac","Type":"ContainerDied","Data":"f400965294df83da970312c01542f42ba08d96b012edb3573480db5c8b7b0480"} Nov 29 00:36:05 crc kubenswrapper[4931]: I1129 00:36:05.807049 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b7v2z" event={"ID":"23f6dae0-af26-46d1-851e-9401172634ac","Type":"ContainerStarted","Data":"d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b"} Nov 29 00:36:05 crc kubenswrapper[4931]: I1129 00:36:05.824362 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b7v2z" podStartSLOduration=2.084165809 podStartE2EDuration="4.824341473s" podCreationTimestamp="2025-11-29 00:36:01 +0000 UTC" firstStartedPulling="2025-11-29 00:36:02.775314873 +0000 UTC m=+2175.937208135" lastFinishedPulling="2025-11-29 00:36:05.515490577 +0000 UTC m=+2178.677383799" observedRunningTime="2025-11-29 00:36:05.821735342 +0000 UTC m=+2178.983628584" watchObservedRunningTime="2025-11-29 00:36:05.824341473 +0000 UTC m=+2178.986234705" Nov 29 00:36:11 crc kubenswrapper[4931]: I1129 00:36:11.523162 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:11 crc kubenswrapper[4931]: I1129 00:36:11.550662 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:11 crc kubenswrapper[4931]: I1129 00:36:11.612353 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:11 crc kubenswrapper[4931]: I1129 00:36:11.949138 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:11 crc kubenswrapper[4931]: I1129 00:36:11.999023 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b7v2z"] Nov 29 00:36:13 crc kubenswrapper[4931]: I1129 00:36:13.902936 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b7v2z" podUID="23f6dae0-af26-46d1-851e-9401172634ac" containerName="registry-server" containerID="cri-o://d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b" gracePeriod=2 Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.489536 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.632525 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f6dae0-af26-46d1-851e-9401172634ac-utilities\") pod \"23f6dae0-af26-46d1-851e-9401172634ac\" (UID: \"23f6dae0-af26-46d1-851e-9401172634ac\") " Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.632572 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f6dae0-af26-46d1-851e-9401172634ac-catalog-content\") pod \"23f6dae0-af26-46d1-851e-9401172634ac\" (UID: \"23f6dae0-af26-46d1-851e-9401172634ac\") " Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.632653 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g757g\" (UniqueName: \"kubernetes.io/projected/23f6dae0-af26-46d1-851e-9401172634ac-kube-api-access-g757g\") pod \"23f6dae0-af26-46d1-851e-9401172634ac\" (UID: \"23f6dae0-af26-46d1-851e-9401172634ac\") " Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.634008 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23f6dae0-af26-46d1-851e-9401172634ac-utilities" (OuterVolumeSpecName: "utilities") pod "23f6dae0-af26-46d1-851e-9401172634ac" (UID: "23f6dae0-af26-46d1-851e-9401172634ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.639755 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23f6dae0-af26-46d1-851e-9401172634ac-kube-api-access-g757g" (OuterVolumeSpecName: "kube-api-access-g757g") pod "23f6dae0-af26-46d1-851e-9401172634ac" (UID: "23f6dae0-af26-46d1-851e-9401172634ac"). InnerVolumeSpecName "kube-api-access-g757g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.709025 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23f6dae0-af26-46d1-851e-9401172634ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23f6dae0-af26-46d1-851e-9401172634ac" (UID: "23f6dae0-af26-46d1-851e-9401172634ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.734415 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f6dae0-af26-46d1-851e-9401172634ac-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.734629 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f6dae0-af26-46d1-851e-9401172634ac-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.734700 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g757g\" (UniqueName: \"kubernetes.io/projected/23f6dae0-af26-46d1-851e-9401172634ac-kube-api-access-g757g\") on node \"crc\" DevicePath \"\"" Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.917723 4931 generic.go:334] "Generic (PLEG): container finished" podID="23f6dae0-af26-46d1-851e-9401172634ac" containerID="d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b" exitCode=0 Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.917835 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b7v2z" event={"ID":"23f6dae0-af26-46d1-851e-9401172634ac","Type":"ContainerDied","Data":"d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b"} Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.917881 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b7v2z" Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.918133 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b7v2z" event={"ID":"23f6dae0-af26-46d1-851e-9401172634ac","Type":"ContainerDied","Data":"abf16e8e46d7bb6591192783e0615846057f232460969efccfc6a17c4230fb37"} Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.918171 4931 scope.go:117] "RemoveContainer" containerID="d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b" Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.947222 4931 scope.go:117] "RemoveContainer" containerID="f400965294df83da970312c01542f42ba08d96b012edb3573480db5c8b7b0480" Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.981183 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b7v2z"] Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.991617 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b7v2z"] Nov 29 00:36:14 crc kubenswrapper[4931]: I1129 00:36:14.994390 4931 scope.go:117] "RemoveContainer" containerID="01d136690bc502ca6ef7b708da2df7f3a468eaddc84aadc4ab46d403e4881f7b" Nov 29 00:36:15 crc kubenswrapper[4931]: I1129 00:36:15.024775 4931 scope.go:117] "RemoveContainer" containerID="d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b" Nov 29 00:36:15 crc kubenswrapper[4931]: E1129 00:36:15.025444 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b\": container with ID starting with d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b not found: ID does not exist" containerID="d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b" Nov 29 00:36:15 crc kubenswrapper[4931]: I1129 00:36:15.025513 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b"} err="failed to get container status \"d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b\": rpc error: code = NotFound desc = could not find container \"d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b\": container with ID starting with d05915ba7c289f85a679cb5bc66c87c7768973f33e397e75d778727786cd9d7b not found: ID does not exist" Nov 29 00:36:15 crc kubenswrapper[4931]: I1129 00:36:15.025542 4931 scope.go:117] "RemoveContainer" containerID="f400965294df83da970312c01542f42ba08d96b012edb3573480db5c8b7b0480" Nov 29 00:36:15 crc kubenswrapper[4931]: E1129 00:36:15.026130 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f400965294df83da970312c01542f42ba08d96b012edb3573480db5c8b7b0480\": container with ID starting with f400965294df83da970312c01542f42ba08d96b012edb3573480db5c8b7b0480 not found: ID does not exist" containerID="f400965294df83da970312c01542f42ba08d96b012edb3573480db5c8b7b0480" Nov 29 00:36:15 crc kubenswrapper[4931]: I1129 00:36:15.026175 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f400965294df83da970312c01542f42ba08d96b012edb3573480db5c8b7b0480"} err="failed to get container status \"f400965294df83da970312c01542f42ba08d96b012edb3573480db5c8b7b0480\": rpc error: code = NotFound desc = could not find container \"f400965294df83da970312c01542f42ba08d96b012edb3573480db5c8b7b0480\": container with ID starting with f400965294df83da970312c01542f42ba08d96b012edb3573480db5c8b7b0480 not found: ID does not exist" Nov 29 00:36:15 crc kubenswrapper[4931]: I1129 00:36:15.026198 4931 scope.go:117] "RemoveContainer" containerID="01d136690bc502ca6ef7b708da2df7f3a468eaddc84aadc4ab46d403e4881f7b" Nov 29 00:36:15 crc kubenswrapper[4931]: E1129 00:36:15.026462 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01d136690bc502ca6ef7b708da2df7f3a468eaddc84aadc4ab46d403e4881f7b\": container with ID starting with 01d136690bc502ca6ef7b708da2df7f3a468eaddc84aadc4ab46d403e4881f7b not found: ID does not exist" containerID="01d136690bc502ca6ef7b708da2df7f3a468eaddc84aadc4ab46d403e4881f7b" Nov 29 00:36:15 crc kubenswrapper[4931]: I1129 00:36:15.026491 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01d136690bc502ca6ef7b708da2df7f3a468eaddc84aadc4ab46d403e4881f7b"} err="failed to get container status \"01d136690bc502ca6ef7b708da2df7f3a468eaddc84aadc4ab46d403e4881f7b\": rpc error: code = NotFound desc = could not find container \"01d136690bc502ca6ef7b708da2df7f3a468eaddc84aadc4ab46d403e4881f7b\": container with ID starting with 01d136690bc502ca6ef7b708da2df7f3a468eaddc84aadc4ab46d403e4881f7b not found: ID does not exist" Nov 29 00:36:15 crc kubenswrapper[4931]: I1129 00:36:15.228125 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23f6dae0-af26-46d1-851e-9401172634ac" path="/var/lib/kubelet/pods/23f6dae0-af26-46d1-851e-9401172634ac/volumes" Nov 29 00:36:23 crc kubenswrapper[4931]: I1129 00:36:23.734998 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:36:23 crc kubenswrapper[4931]: I1129 00:36:23.735771 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:36:23 crc kubenswrapper[4931]: I1129 00:36:23.735859 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:36:23 crc kubenswrapper[4931]: I1129 00:36:23.737004 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc12cbef83d264cf45b46aedbbb0832fdfe3ed670622f291973be8df8b6edd8a"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 00:36:23 crc kubenswrapper[4931]: I1129 00:36:23.737097 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://bc12cbef83d264cf45b46aedbbb0832fdfe3ed670622f291973be8df8b6edd8a" gracePeriod=600 Nov 29 00:36:24 crc kubenswrapper[4931]: I1129 00:36:24.015285 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="bc12cbef83d264cf45b46aedbbb0832fdfe3ed670622f291973be8df8b6edd8a" exitCode=0 Nov 29 00:36:24 crc kubenswrapper[4931]: I1129 00:36:24.015537 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"bc12cbef83d264cf45b46aedbbb0832fdfe3ed670622f291973be8df8b6edd8a"} Nov 29 00:36:24 crc kubenswrapper[4931]: I1129 00:36:24.016272 4931 scope.go:117] "RemoveContainer" containerID="c16e37c0cc656d16693fb9c1284ed347cdc4e5353176e9222909c671b08aa357" Nov 29 00:36:25 crc kubenswrapper[4931]: I1129 00:36:25.035652 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2"} Nov 29 00:36:26 crc kubenswrapper[4931]: I1129 00:36:26.048056 4931 generic.go:334] "Generic (PLEG): container finished" podID="5da7ac90-6a60-4473-b8b8-5531d4c2ac68" containerID="a2a604325fc7447492aff40493aa7521f88a5850584df6c4ce79e0309ff3d7b3" exitCode=0 Nov 29 00:36:26 crc kubenswrapper[4931]: I1129 00:36:26.048096 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" event={"ID":"5da7ac90-6a60-4473-b8b8-5531d4c2ac68","Type":"ContainerDied","Data":"a2a604325fc7447492aff40493aa7521f88a5850584df6c4ce79e0309ff3d7b3"} Nov 29 00:36:27 crc kubenswrapper[4931]: I1129 00:36:27.507045 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:36:27 crc kubenswrapper[4931]: I1129 00:36:27.590672 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-inventory\") pod \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\" (UID: \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\") " Nov 29 00:36:27 crc kubenswrapper[4931]: I1129 00:36:27.590732 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-ssh-key\") pod \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\" (UID: \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\") " Nov 29 00:36:27 crc kubenswrapper[4931]: I1129 00:36:27.590937 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjrjv\" (UniqueName: \"kubernetes.io/projected/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-kube-api-access-bjrjv\") pod \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\" (UID: \"5da7ac90-6a60-4473-b8b8-5531d4c2ac68\") " Nov 29 00:36:27 crc kubenswrapper[4931]: I1129 00:36:27.596259 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-kube-api-access-bjrjv" (OuterVolumeSpecName: "kube-api-access-bjrjv") pod "5da7ac90-6a60-4473-b8b8-5531d4c2ac68" (UID: "5da7ac90-6a60-4473-b8b8-5531d4c2ac68"). InnerVolumeSpecName "kube-api-access-bjrjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:36:27 crc kubenswrapper[4931]: I1129 00:36:27.633270 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5da7ac90-6a60-4473-b8b8-5531d4c2ac68" (UID: "5da7ac90-6a60-4473-b8b8-5531d4c2ac68"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:36:27 crc kubenswrapper[4931]: I1129 00:36:27.633534 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-inventory" (OuterVolumeSpecName: "inventory") pod "5da7ac90-6a60-4473-b8b8-5531d4c2ac68" (UID: "5da7ac90-6a60-4473-b8b8-5531d4c2ac68"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:36:27 crc kubenswrapper[4931]: I1129 00:36:27.693568 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjrjv\" (UniqueName: \"kubernetes.io/projected/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-kube-api-access-bjrjv\") on node \"crc\" DevicePath \"\"" Nov 29 00:36:27 crc kubenswrapper[4931]: I1129 00:36:27.693600 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:36:27 crc kubenswrapper[4931]: I1129 00:36:27.693608 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5da7ac90-6a60-4473-b8b8-5531d4c2ac68-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.070949 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" event={"ID":"5da7ac90-6a60-4473-b8b8-5531d4c2ac68","Type":"ContainerDied","Data":"8ef65c183c01bac36c421c67fbe0f286f01a224b692b227269942559241e63e1"} Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.071265 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ef65c183c01bac36c421c67fbe0f286f01a224b692b227269942559241e63e1" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.071022 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-brwbq" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.146698 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-gfsvs"] Nov 29 00:36:28 crc kubenswrapper[4931]: E1129 00:36:28.147161 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da7ac90-6a60-4473-b8b8-5531d4c2ac68" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.147187 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da7ac90-6a60-4473-b8b8-5531d4c2ac68" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 29 00:36:28 crc kubenswrapper[4931]: E1129 00:36:28.147217 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f6dae0-af26-46d1-851e-9401172634ac" containerName="extract-utilities" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.147226 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f6dae0-af26-46d1-851e-9401172634ac" containerName="extract-utilities" Nov 29 00:36:28 crc kubenswrapper[4931]: E1129 00:36:28.147290 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f6dae0-af26-46d1-851e-9401172634ac" containerName="extract-content" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.147302 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f6dae0-af26-46d1-851e-9401172634ac" containerName="extract-content" Nov 29 00:36:28 crc kubenswrapper[4931]: E1129 00:36:28.147316 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f6dae0-af26-46d1-851e-9401172634ac" containerName="registry-server" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.147324 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f6dae0-af26-46d1-851e-9401172634ac" containerName="registry-server" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.147714 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="5da7ac90-6a60-4473-b8b8-5531d4c2ac68" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.147749 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="23f6dae0-af26-46d1-851e-9401172634ac" containerName="registry-server" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.149030 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.155432 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.155789 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.156738 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.164184 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.183801 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-gfsvs"] Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.304767 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j6jm\" (UniqueName: \"kubernetes.io/projected/333e0546-64b0-412e-9059-1e7a4082d360-kube-api-access-5j6jm\") pod \"ssh-known-hosts-edpm-deployment-gfsvs\" (UID: \"333e0546-64b0-412e-9059-1e7a4082d360\") " pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.304859 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/333e0546-64b0-412e-9059-1e7a4082d360-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-gfsvs\" (UID: \"333e0546-64b0-412e-9059-1e7a4082d360\") " pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.304887 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/333e0546-64b0-412e-9059-1e7a4082d360-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-gfsvs\" (UID: \"333e0546-64b0-412e-9059-1e7a4082d360\") " pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.406926 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j6jm\" (UniqueName: \"kubernetes.io/projected/333e0546-64b0-412e-9059-1e7a4082d360-kube-api-access-5j6jm\") pod \"ssh-known-hosts-edpm-deployment-gfsvs\" (UID: \"333e0546-64b0-412e-9059-1e7a4082d360\") " pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.406982 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/333e0546-64b0-412e-9059-1e7a4082d360-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-gfsvs\" (UID: \"333e0546-64b0-412e-9059-1e7a4082d360\") " pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.407004 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/333e0546-64b0-412e-9059-1e7a4082d360-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-gfsvs\" (UID: \"333e0546-64b0-412e-9059-1e7a4082d360\") " pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.415591 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/333e0546-64b0-412e-9059-1e7a4082d360-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-gfsvs\" (UID: \"333e0546-64b0-412e-9059-1e7a4082d360\") " pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.415633 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/333e0546-64b0-412e-9059-1e7a4082d360-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-gfsvs\" (UID: \"333e0546-64b0-412e-9059-1e7a4082d360\") " pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.427286 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j6jm\" (UniqueName: \"kubernetes.io/projected/333e0546-64b0-412e-9059-1e7a4082d360-kube-api-access-5j6jm\") pod \"ssh-known-hosts-edpm-deployment-gfsvs\" (UID: \"333e0546-64b0-412e-9059-1e7a4082d360\") " pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:28 crc kubenswrapper[4931]: I1129 00:36:28.484194 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:29 crc kubenswrapper[4931]: I1129 00:36:29.052306 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-gfsvs"] Nov 29 00:36:29 crc kubenswrapper[4931]: I1129 00:36:29.080374 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" event={"ID":"333e0546-64b0-412e-9059-1e7a4082d360","Type":"ContainerStarted","Data":"a67fc4b8c99e8e0c30cf443ad81fb2b92ac5cb7324e265b6c439723dc9a8844e"} Nov 29 00:36:30 crc kubenswrapper[4931]: I1129 00:36:30.092990 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" event={"ID":"333e0546-64b0-412e-9059-1e7a4082d360","Type":"ContainerStarted","Data":"871da4a179a76b163eb1c545c07e3fa153d3b0bba4a07940ec993e9919402d8d"} Nov 29 00:36:30 crc kubenswrapper[4931]: I1129 00:36:30.128137 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" podStartSLOduration=1.5667878160000002 podStartE2EDuration="2.128117564s" podCreationTimestamp="2025-11-29 00:36:28 +0000 UTC" firstStartedPulling="2025-11-29 00:36:29.061224147 +0000 UTC m=+2202.223117379" lastFinishedPulling="2025-11-29 00:36:29.622553895 +0000 UTC m=+2202.784447127" observedRunningTime="2025-11-29 00:36:30.116236763 +0000 UTC m=+2203.278130025" watchObservedRunningTime="2025-11-29 00:36:30.128117564 +0000 UTC m=+2203.290010806" Nov 29 00:36:37 crc kubenswrapper[4931]: I1129 00:36:37.168930 4931 generic.go:334] "Generic (PLEG): container finished" podID="333e0546-64b0-412e-9059-1e7a4082d360" containerID="871da4a179a76b163eb1c545c07e3fa153d3b0bba4a07940ec993e9919402d8d" exitCode=0 Nov 29 00:36:37 crc kubenswrapper[4931]: I1129 00:36:37.169090 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" event={"ID":"333e0546-64b0-412e-9059-1e7a4082d360","Type":"ContainerDied","Data":"871da4a179a76b163eb1c545c07e3fa153d3b0bba4a07940ec993e9919402d8d"} Nov 29 00:36:38 crc kubenswrapper[4931]: I1129 00:36:38.694429 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:38 crc kubenswrapper[4931]: I1129 00:36:38.734418 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5j6jm\" (UniqueName: \"kubernetes.io/projected/333e0546-64b0-412e-9059-1e7a4082d360-kube-api-access-5j6jm\") pod \"333e0546-64b0-412e-9059-1e7a4082d360\" (UID: \"333e0546-64b0-412e-9059-1e7a4082d360\") " Nov 29 00:36:38 crc kubenswrapper[4931]: I1129 00:36:38.734544 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/333e0546-64b0-412e-9059-1e7a4082d360-ssh-key-openstack-edpm-ipam\") pod \"333e0546-64b0-412e-9059-1e7a4082d360\" (UID: \"333e0546-64b0-412e-9059-1e7a4082d360\") " Nov 29 00:36:38 crc kubenswrapper[4931]: I1129 00:36:38.734592 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/333e0546-64b0-412e-9059-1e7a4082d360-inventory-0\") pod \"333e0546-64b0-412e-9059-1e7a4082d360\" (UID: \"333e0546-64b0-412e-9059-1e7a4082d360\") " Nov 29 00:36:38 crc kubenswrapper[4931]: I1129 00:36:38.746203 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/333e0546-64b0-412e-9059-1e7a4082d360-kube-api-access-5j6jm" (OuterVolumeSpecName: "kube-api-access-5j6jm") pod "333e0546-64b0-412e-9059-1e7a4082d360" (UID: "333e0546-64b0-412e-9059-1e7a4082d360"). InnerVolumeSpecName "kube-api-access-5j6jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:36:38 crc kubenswrapper[4931]: I1129 00:36:38.773572 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/333e0546-64b0-412e-9059-1e7a4082d360-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "333e0546-64b0-412e-9059-1e7a4082d360" (UID: "333e0546-64b0-412e-9059-1e7a4082d360"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:36:38 crc kubenswrapper[4931]: I1129 00:36:38.775972 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/333e0546-64b0-412e-9059-1e7a4082d360-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "333e0546-64b0-412e-9059-1e7a4082d360" (UID: "333e0546-64b0-412e-9059-1e7a4082d360"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:36:38 crc kubenswrapper[4931]: I1129 00:36:38.837397 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5j6jm\" (UniqueName: \"kubernetes.io/projected/333e0546-64b0-412e-9059-1e7a4082d360-kube-api-access-5j6jm\") on node \"crc\" DevicePath \"\"" Nov 29 00:36:38 crc kubenswrapper[4931]: I1129 00:36:38.837439 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/333e0546-64b0-412e-9059-1e7a4082d360-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 29 00:36:38 crc kubenswrapper[4931]: I1129 00:36:38.837453 4931 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/333e0546-64b0-412e-9059-1e7a4082d360-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.190538 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" event={"ID":"333e0546-64b0-412e-9059-1e7a4082d360","Type":"ContainerDied","Data":"a67fc4b8c99e8e0c30cf443ad81fb2b92ac5cb7324e265b6c439723dc9a8844e"} Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.190599 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a67fc4b8c99e8e0c30cf443ad81fb2b92ac5cb7324e265b6c439723dc9a8844e" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.190628 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-gfsvs" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.267502 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp"] Nov 29 00:36:39 crc kubenswrapper[4931]: E1129 00:36:39.267987 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333e0546-64b0-412e-9059-1e7a4082d360" containerName="ssh-known-hosts-edpm-deployment" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.268012 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="333e0546-64b0-412e-9059-1e7a4082d360" containerName="ssh-known-hosts-edpm-deployment" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.268259 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="333e0546-64b0-412e-9059-1e7a4082d360" containerName="ssh-known-hosts-edpm-deployment" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.269040 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.272776 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.274132 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.274739 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.285362 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp"] Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.287272 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.348823 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkqcv\" (UniqueName: \"kubernetes.io/projected/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-kube-api-access-rkqcv\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9gcqp\" (UID: \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.348938 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9gcqp\" (UID: \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.348970 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9gcqp\" (UID: \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.450854 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkqcv\" (UniqueName: \"kubernetes.io/projected/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-kube-api-access-rkqcv\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9gcqp\" (UID: \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.451250 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9gcqp\" (UID: \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.451397 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9gcqp\" (UID: \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.456621 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9gcqp\" (UID: \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.458365 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9gcqp\" (UID: \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.471721 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkqcv\" (UniqueName: \"kubernetes.io/projected/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-kube-api-access-rkqcv\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9gcqp\" (UID: \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:39 crc kubenswrapper[4931]: I1129 00:36:39.595105 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:40 crc kubenswrapper[4931]: I1129 00:36:40.185217 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp"] Nov 29 00:36:40 crc kubenswrapper[4931]: I1129 00:36:40.200468 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" event={"ID":"6c31a7d4-ea5c-4434-a405-cd4148ca5a87","Type":"ContainerStarted","Data":"1694100122da904fc9b53b32d17de49f8dbc6b17c4366906d0a83ae073519941"} Nov 29 00:36:41 crc kubenswrapper[4931]: I1129 00:36:41.210978 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" event={"ID":"6c31a7d4-ea5c-4434-a405-cd4148ca5a87","Type":"ContainerStarted","Data":"1c2d5478d9a6cb8d465e0bf600f151c3edb34f9747231f6480e0d34d000ccb4b"} Nov 29 00:36:41 crc kubenswrapper[4931]: I1129 00:36:41.235345 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" podStartSLOduration=1.697562198 podStartE2EDuration="2.23532036s" podCreationTimestamp="2025-11-29 00:36:39 +0000 UTC" firstStartedPulling="2025-11-29 00:36:40.188659753 +0000 UTC m=+2213.350553015" lastFinishedPulling="2025-11-29 00:36:40.726417915 +0000 UTC m=+2213.888311177" observedRunningTime="2025-11-29 00:36:41.230368788 +0000 UTC m=+2214.392262060" watchObservedRunningTime="2025-11-29 00:36:41.23532036 +0000 UTC m=+2214.397213602" Nov 29 00:36:50 crc kubenswrapper[4931]: I1129 00:36:50.315789 4931 generic.go:334] "Generic (PLEG): container finished" podID="6c31a7d4-ea5c-4434-a405-cd4148ca5a87" containerID="1c2d5478d9a6cb8d465e0bf600f151c3edb34f9747231f6480e0d34d000ccb4b" exitCode=0 Nov 29 00:36:50 crc kubenswrapper[4931]: I1129 00:36:50.315913 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" event={"ID":"6c31a7d4-ea5c-4434-a405-cd4148ca5a87","Type":"ContainerDied","Data":"1c2d5478d9a6cb8d465e0bf600f151c3edb34f9747231f6480e0d34d000ccb4b"} Nov 29 00:36:51 crc kubenswrapper[4931]: I1129 00:36:51.807259 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:51 crc kubenswrapper[4931]: I1129 00:36:51.896832 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-inventory\") pod \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\" (UID: \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\") " Nov 29 00:36:51 crc kubenswrapper[4931]: I1129 00:36:51.896936 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkqcv\" (UniqueName: \"kubernetes.io/projected/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-kube-api-access-rkqcv\") pod \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\" (UID: \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\") " Nov 29 00:36:51 crc kubenswrapper[4931]: I1129 00:36:51.896995 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-ssh-key\") pod \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\" (UID: \"6c31a7d4-ea5c-4434-a405-cd4148ca5a87\") " Nov 29 00:36:51 crc kubenswrapper[4931]: I1129 00:36:51.925632 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-kube-api-access-rkqcv" (OuterVolumeSpecName: "kube-api-access-rkqcv") pod "6c31a7d4-ea5c-4434-a405-cd4148ca5a87" (UID: "6c31a7d4-ea5c-4434-a405-cd4148ca5a87"). InnerVolumeSpecName "kube-api-access-rkqcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:36:51 crc kubenswrapper[4931]: I1129 00:36:51.960417 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6c31a7d4-ea5c-4434-a405-cd4148ca5a87" (UID: "6c31a7d4-ea5c-4434-a405-cd4148ca5a87"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:36:51 crc kubenswrapper[4931]: I1129 00:36:51.967799 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-inventory" (OuterVolumeSpecName: "inventory") pod "6c31a7d4-ea5c-4434-a405-cd4148ca5a87" (UID: "6c31a7d4-ea5c-4434-a405-cd4148ca5a87"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:36:51 crc kubenswrapper[4931]: I1129 00:36:51.998768 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:36:51 crc kubenswrapper[4931]: I1129 00:36:51.998800 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkqcv\" (UniqueName: \"kubernetes.io/projected/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-kube-api-access-rkqcv\") on node \"crc\" DevicePath \"\"" Nov 29 00:36:51 crc kubenswrapper[4931]: I1129 00:36:51.998830 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c31a7d4-ea5c-4434-a405-cd4148ca5a87-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.346549 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" event={"ID":"6c31a7d4-ea5c-4434-a405-cd4148ca5a87","Type":"ContainerDied","Data":"1694100122da904fc9b53b32d17de49f8dbc6b17c4366906d0a83ae073519941"} Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.346591 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1694100122da904fc9b53b32d17de49f8dbc6b17c4366906d0a83ae073519941" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.346618 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9gcqp" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.430494 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs"] Nov 29 00:36:52 crc kubenswrapper[4931]: E1129 00:36:52.430940 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c31a7d4-ea5c-4434-a405-cd4148ca5a87" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.430961 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c31a7d4-ea5c-4434-a405-cd4148ca5a87" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.431220 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c31a7d4-ea5c-4434-a405-cd4148ca5a87" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.431989 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.434576 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.434608 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.435292 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.435461 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.444395 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs"] Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.512050 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs\" (UID: \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.512137 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkrmj\" (UniqueName: \"kubernetes.io/projected/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-kube-api-access-tkrmj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs\" (UID: \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.512182 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs\" (UID: \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.614377 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs\" (UID: \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.614506 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkrmj\" (UniqueName: \"kubernetes.io/projected/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-kube-api-access-tkrmj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs\" (UID: \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.614584 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs\" (UID: \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.622932 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs\" (UID: \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.624991 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs\" (UID: \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.635290 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkrmj\" (UniqueName: \"kubernetes.io/projected/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-kube-api-access-tkrmj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs\" (UID: \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:36:52 crc kubenswrapper[4931]: I1129 00:36:52.757419 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:36:53 crc kubenswrapper[4931]: W1129 00:36:53.334872 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod417e828a_2ea5_4ad0_86e0_b5d612cccbe2.slice/crio-e95b758dd432d4c9410c2f7501086db6d6e9554f9795ac42d61cbef6abeedf0d WatchSource:0}: Error finding container e95b758dd432d4c9410c2f7501086db6d6e9554f9795ac42d61cbef6abeedf0d: Status 404 returned error can't find the container with id e95b758dd432d4c9410c2f7501086db6d6e9554f9795ac42d61cbef6abeedf0d Nov 29 00:36:53 crc kubenswrapper[4931]: I1129 00:36:53.339745 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs"] Nov 29 00:36:53 crc kubenswrapper[4931]: I1129 00:36:53.356618 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" event={"ID":"417e828a-2ea5-4ad0-86e0-b5d612cccbe2","Type":"ContainerStarted","Data":"e95b758dd432d4c9410c2f7501086db6d6e9554f9795ac42d61cbef6abeedf0d"} Nov 29 00:36:54 crc kubenswrapper[4931]: I1129 00:36:54.371022 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" event={"ID":"417e828a-2ea5-4ad0-86e0-b5d612cccbe2","Type":"ContainerStarted","Data":"5ad8f55be78fe5eb4c61a2eac6b5610e0d5dd9475be1b4fbee8159bf6a34d48c"} Nov 29 00:36:54 crc kubenswrapper[4931]: I1129 00:36:54.401515 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" podStartSLOduration=1.911473321 podStartE2EDuration="2.401486943s" podCreationTimestamp="2025-11-29 00:36:52 +0000 UTC" firstStartedPulling="2025-11-29 00:36:53.338491648 +0000 UTC m=+2226.500384880" lastFinishedPulling="2025-11-29 00:36:53.82850527 +0000 UTC m=+2226.990398502" observedRunningTime="2025-11-29 00:36:54.386924425 +0000 UTC m=+2227.548817697" watchObservedRunningTime="2025-11-29 00:36:54.401486943 +0000 UTC m=+2227.563380215" Nov 29 00:37:04 crc kubenswrapper[4931]: I1129 00:37:04.470868 4931 generic.go:334] "Generic (PLEG): container finished" podID="417e828a-2ea5-4ad0-86e0-b5d612cccbe2" containerID="5ad8f55be78fe5eb4c61a2eac6b5610e0d5dd9475be1b4fbee8159bf6a34d48c" exitCode=0 Nov 29 00:37:04 crc kubenswrapper[4931]: I1129 00:37:04.470963 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" event={"ID":"417e828a-2ea5-4ad0-86e0-b5d612cccbe2","Type":"ContainerDied","Data":"5ad8f55be78fe5eb4c61a2eac6b5610e0d5dd9475be1b4fbee8159bf6a34d48c"} Nov 29 00:37:05 crc kubenswrapper[4931]: I1129 00:37:05.929119 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.094953 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-inventory\") pod \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\" (UID: \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\") " Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.095236 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkrmj\" (UniqueName: \"kubernetes.io/projected/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-kube-api-access-tkrmj\") pod \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\" (UID: \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\") " Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.095283 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-ssh-key\") pod \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\" (UID: \"417e828a-2ea5-4ad0-86e0-b5d612cccbe2\") " Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.104872 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-kube-api-access-tkrmj" (OuterVolumeSpecName: "kube-api-access-tkrmj") pod "417e828a-2ea5-4ad0-86e0-b5d612cccbe2" (UID: "417e828a-2ea5-4ad0-86e0-b5d612cccbe2"). InnerVolumeSpecName "kube-api-access-tkrmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.148131 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "417e828a-2ea5-4ad0-86e0-b5d612cccbe2" (UID: "417e828a-2ea5-4ad0-86e0-b5d612cccbe2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.155223 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-inventory" (OuterVolumeSpecName: "inventory") pod "417e828a-2ea5-4ad0-86e0-b5d612cccbe2" (UID: "417e828a-2ea5-4ad0-86e0-b5d612cccbe2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.197365 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkrmj\" (UniqueName: \"kubernetes.io/projected/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-kube-api-access-tkrmj\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.197403 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.197416 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/417e828a-2ea5-4ad0-86e0-b5d612cccbe2-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.495423 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" event={"ID":"417e828a-2ea5-4ad0-86e0-b5d612cccbe2","Type":"ContainerDied","Data":"e95b758dd432d4c9410c2f7501086db6d6e9554f9795ac42d61cbef6abeedf0d"} Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.495476 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e95b758dd432d4c9410c2f7501086db6d6e9554f9795ac42d61cbef6abeedf0d" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.495569 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.600543 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7"] Nov 29 00:37:06 crc kubenswrapper[4931]: E1129 00:37:06.601269 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="417e828a-2ea5-4ad0-86e0-b5d612cccbe2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.601315 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="417e828a-2ea5-4ad0-86e0-b5d612cccbe2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.601842 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="417e828a-2ea5-4ad0-86e0-b5d612cccbe2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.603260 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.613755 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7"] Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.614318 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.614720 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.614900 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.615016 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.615132 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.615246 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.615713 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.615838 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.707696 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.709267 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.709405 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.709484 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.709542 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.709692 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.710762 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.710919 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.711055 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.711188 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.711260 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.711418 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqp2p\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-kube-api-access-vqp2p\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.711498 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.711547 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814225 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814329 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814394 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814451 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814504 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814540 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814578 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814668 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814719 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814768 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814830 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814917 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqp2p\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-kube-api-access-vqp2p\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.814973 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.815010 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.826094 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.826149 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.826635 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.826761 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.827068 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.827217 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.827354 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.827680 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.827784 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.828924 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.829456 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.834790 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.837451 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.850183 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqp2p\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-kube-api-access-vqp2p\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:06 crc kubenswrapper[4931]: I1129 00:37:06.941192 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:07 crc kubenswrapper[4931]: I1129 00:37:07.526854 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7"] Nov 29 00:37:08 crc kubenswrapper[4931]: I1129 00:37:08.516701 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" event={"ID":"88f61c01-c5da-4d84-a484-48f4e3b4e3e4","Type":"ContainerStarted","Data":"05041930cbaaee6a5d836b48fa84ff01d30224f75bdcf872586679c385706a92"} Nov 29 00:37:10 crc kubenswrapper[4931]: I1129 00:37:10.542011 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" event={"ID":"88f61c01-c5da-4d84-a484-48f4e3b4e3e4","Type":"ContainerStarted","Data":"a277a04f49feef29e6fb5e77e43ec853ab7895f7db9fe8a83304491cbc78206d"} Nov 29 00:37:10 crc kubenswrapper[4931]: I1129 00:37:10.574968 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" podStartSLOduration=2.302578641 podStartE2EDuration="4.5749447s" podCreationTimestamp="2025-11-29 00:37:06 +0000 UTC" firstStartedPulling="2025-11-29 00:37:07.532130917 +0000 UTC m=+2240.694024149" lastFinishedPulling="2025-11-29 00:37:09.804496966 +0000 UTC m=+2242.966390208" observedRunningTime="2025-11-29 00:37:10.568092073 +0000 UTC m=+2243.729985325" watchObservedRunningTime="2025-11-29 00:37:10.5749447 +0000 UTC m=+2243.736837962" Nov 29 00:37:53 crc kubenswrapper[4931]: I1129 00:37:53.060683 4931 generic.go:334] "Generic (PLEG): container finished" podID="88f61c01-c5da-4d84-a484-48f4e3b4e3e4" containerID="a277a04f49feef29e6fb5e77e43ec853ab7895f7db9fe8a83304491cbc78206d" exitCode=0 Nov 29 00:37:53 crc kubenswrapper[4931]: I1129 00:37:53.060777 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" event={"ID":"88f61c01-c5da-4d84-a484-48f4e3b4e3e4","Type":"ContainerDied","Data":"a277a04f49feef29e6fb5e77e43ec853ab7895f7db9fe8a83304491cbc78206d"} Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.527199 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.557515 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-neutron-metadata-combined-ca-bundle\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.557586 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-inventory\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.557697 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-ovn-default-certs-0\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.557746 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-repo-setup-combined-ca-bundle\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.557797 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.557858 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqp2p\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-kube-api-access-vqp2p\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.557889 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-libvirt-combined-ca-bundle\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.557933 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-bootstrap-combined-ca-bundle\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.557960 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.557987 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-telemetry-combined-ca-bundle\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.558032 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-ssh-key\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.558048 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-ovn-combined-ca-bundle\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.558085 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.558102 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-nova-combined-ca-bundle\") pod \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\" (UID: \"88f61c01-c5da-4d84-a484-48f4e3b4e3e4\") " Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.567199 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.567632 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.567723 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.568564 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.569043 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.569624 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.570026 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.572278 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.574176 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.574276 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-kube-api-access-vqp2p" (OuterVolumeSpecName: "kube-api-access-vqp2p") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "kube-api-access-vqp2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.574978 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.576597 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.593732 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-inventory" (OuterVolumeSpecName: "inventory") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.595315 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "88f61c01-c5da-4d84-a484-48f4e3b4e3e4" (UID: "88f61c01-c5da-4d84-a484-48f4e3b4e3e4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661018 4931 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661094 4931 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661142 4931 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661158 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqp2p\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-kube-api-access-vqp2p\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661171 4931 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661182 4931 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661223 4931 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661239 4931 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661252 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661264 4931 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661304 4931 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661320 4931 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661333 4931 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:54 crc kubenswrapper[4931]: I1129 00:37:54.661344 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88f61c01-c5da-4d84-a484-48f4e3b4e3e4-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.080880 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" event={"ID":"88f61c01-c5da-4d84-a484-48f4e3b4e3e4","Type":"ContainerDied","Data":"05041930cbaaee6a5d836b48fa84ff01d30224f75bdcf872586679c385706a92"} Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.080928 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05041930cbaaee6a5d836b48fa84ff01d30224f75bdcf872586679c385706a92" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.081017 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.230686 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv"] Nov 29 00:37:55 crc kubenswrapper[4931]: E1129 00:37:55.231266 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88f61c01-c5da-4d84-a484-48f4e3b4e3e4" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.231298 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="88f61c01-c5da-4d84-a484-48f4e3b4e3e4" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.231632 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="88f61c01-c5da-4d84-a484-48f4e3b4e3e4" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.232687 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.236033 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.236240 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.236309 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.239188 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.240047 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.247386 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv"] Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.274149 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnnwh\" (UniqueName: \"kubernetes.io/projected/7715e2f2-11a1-4702-86f0-84744ae25c5d-kube-api-access-jnnwh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.274318 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7715e2f2-11a1-4702-86f0-84744ae25c5d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.274374 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.274407 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.274481 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.376575 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnnwh\" (UniqueName: \"kubernetes.io/projected/7715e2f2-11a1-4702-86f0-84744ae25c5d-kube-api-access-jnnwh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.377246 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7715e2f2-11a1-4702-86f0-84744ae25c5d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.379190 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7715e2f2-11a1-4702-86f0-84744ae25c5d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.379280 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.379349 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.379462 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.386635 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.386653 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.387077 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.396291 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnnwh\" (UniqueName: \"kubernetes.io/projected/7715e2f2-11a1-4702-86f0-84744ae25c5d-kube-api-access-jnnwh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4p9cv\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:55 crc kubenswrapper[4931]: I1129 00:37:55.560082 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:37:56 crc kubenswrapper[4931]: I1129 00:37:56.184423 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv"] Nov 29 00:37:56 crc kubenswrapper[4931]: I1129 00:37:56.191091 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 00:37:57 crc kubenswrapper[4931]: I1129 00:37:57.116220 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" event={"ID":"7715e2f2-11a1-4702-86f0-84744ae25c5d","Type":"ContainerStarted","Data":"da76e52811a9371cf3e4832c2550ee3930e6da222fabd9a4a848ce8de29e8f73"} Nov 29 00:37:58 crc kubenswrapper[4931]: I1129 00:37:58.129572 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" event={"ID":"7715e2f2-11a1-4702-86f0-84744ae25c5d","Type":"ContainerStarted","Data":"de640fc71eede51f84152f208e33ff6104ff383f58197d16efd8df93c4b9d630"} Nov 29 00:37:58 crc kubenswrapper[4931]: I1129 00:37:58.159596 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" podStartSLOduration=2.514554261 podStartE2EDuration="3.159579988s" podCreationTimestamp="2025-11-29 00:37:55 +0000 UTC" firstStartedPulling="2025-11-29 00:37:56.190661873 +0000 UTC m=+2289.352555145" lastFinishedPulling="2025-11-29 00:37:56.8356876 +0000 UTC m=+2289.997580872" observedRunningTime="2025-11-29 00:37:58.154220224 +0000 UTC m=+2291.316113536" watchObservedRunningTime="2025-11-29 00:37:58.159579988 +0000 UTC m=+2291.321473220" Nov 29 00:38:53 crc kubenswrapper[4931]: I1129 00:38:53.734603 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:38:53 crc kubenswrapper[4931]: I1129 00:38:53.735157 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:39:07 crc kubenswrapper[4931]: I1129 00:39:07.890749 4931 generic.go:334] "Generic (PLEG): container finished" podID="7715e2f2-11a1-4702-86f0-84744ae25c5d" containerID="de640fc71eede51f84152f208e33ff6104ff383f58197d16efd8df93c4b9d630" exitCode=0 Nov 29 00:39:07 crc kubenswrapper[4931]: I1129 00:39:07.890844 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" event={"ID":"7715e2f2-11a1-4702-86f0-84744ae25c5d","Type":"ContainerDied","Data":"de640fc71eede51f84152f208e33ff6104ff383f58197d16efd8df93c4b9d630"} Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.306622 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.509410 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-ovn-combined-ca-bundle\") pod \"7715e2f2-11a1-4702-86f0-84744ae25c5d\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.509449 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnnwh\" (UniqueName: \"kubernetes.io/projected/7715e2f2-11a1-4702-86f0-84744ae25c5d-kube-api-access-jnnwh\") pod \"7715e2f2-11a1-4702-86f0-84744ae25c5d\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.509530 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-inventory\") pod \"7715e2f2-11a1-4702-86f0-84744ae25c5d\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.509582 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-ssh-key\") pod \"7715e2f2-11a1-4702-86f0-84744ae25c5d\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.509602 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7715e2f2-11a1-4702-86f0-84744ae25c5d-ovncontroller-config-0\") pod \"7715e2f2-11a1-4702-86f0-84744ae25c5d\" (UID: \"7715e2f2-11a1-4702-86f0-84744ae25c5d\") " Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.514550 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7715e2f2-11a1-4702-86f0-84744ae25c5d-kube-api-access-jnnwh" (OuterVolumeSpecName: "kube-api-access-jnnwh") pod "7715e2f2-11a1-4702-86f0-84744ae25c5d" (UID: "7715e2f2-11a1-4702-86f0-84744ae25c5d"). InnerVolumeSpecName "kube-api-access-jnnwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.515717 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "7715e2f2-11a1-4702-86f0-84744ae25c5d" (UID: "7715e2f2-11a1-4702-86f0-84744ae25c5d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.534603 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-inventory" (OuterVolumeSpecName: "inventory") pod "7715e2f2-11a1-4702-86f0-84744ae25c5d" (UID: "7715e2f2-11a1-4702-86f0-84744ae25c5d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.537768 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7715e2f2-11a1-4702-86f0-84744ae25c5d-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "7715e2f2-11a1-4702-86f0-84744ae25c5d" (UID: "7715e2f2-11a1-4702-86f0-84744ae25c5d"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.543116 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7715e2f2-11a1-4702-86f0-84744ae25c5d" (UID: "7715e2f2-11a1-4702-86f0-84744ae25c5d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.611360 4931 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.611397 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnnwh\" (UniqueName: \"kubernetes.io/projected/7715e2f2-11a1-4702-86f0-84744ae25c5d-kube-api-access-jnnwh\") on node \"crc\" DevicePath \"\"" Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.611407 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.611416 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7715e2f2-11a1-4702-86f0-84744ae25c5d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.611425 4931 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7715e2f2-11a1-4702-86f0-84744ae25c5d-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.913675 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" event={"ID":"7715e2f2-11a1-4702-86f0-84744ae25c5d","Type":"ContainerDied","Data":"da76e52811a9371cf3e4832c2550ee3930e6da222fabd9a4a848ce8de29e8f73"} Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.913722 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da76e52811a9371cf3e4832c2550ee3930e6da222fabd9a4a848ce8de29e8f73" Nov 29 00:39:09 crc kubenswrapper[4931]: I1129 00:39:09.913785 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4p9cv" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.013402 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52"] Nov 29 00:39:10 crc kubenswrapper[4931]: E1129 00:39:10.013904 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7715e2f2-11a1-4702-86f0-84744ae25c5d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.013927 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="7715e2f2-11a1-4702-86f0-84744ae25c5d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.014146 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="7715e2f2-11a1-4702-86f0-84744ae25c5d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.014939 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.017599 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.017621 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.017980 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.018004 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.018378 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.022557 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.031228 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52"] Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.121363 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.121427 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.121544 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42lnk\" (UniqueName: \"kubernetes.io/projected/054b5e53-28bc-4222-9f59-ad0ebd2cf767-kube-api-access-42lnk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.121592 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.121776 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.121957 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.224337 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.224433 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.224516 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42lnk\" (UniqueName: \"kubernetes.io/projected/054b5e53-28bc-4222-9f59-ad0ebd2cf767-kube-api-access-42lnk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.224555 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.224625 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.224668 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.228281 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.228365 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.229550 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.231091 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.232616 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.258844 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42lnk\" (UniqueName: \"kubernetes.io/projected/054b5e53-28bc-4222-9f59-ad0ebd2cf767-kube-api-access-42lnk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:10 crc kubenswrapper[4931]: I1129 00:39:10.333666 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:39:11 crc kubenswrapper[4931]: I1129 00:39:11.044594 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52"] Nov 29 00:39:11 crc kubenswrapper[4931]: I1129 00:39:11.976983 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" event={"ID":"054b5e53-28bc-4222-9f59-ad0ebd2cf767","Type":"ContainerStarted","Data":"b2becb29990a8278a6fe5aacf594b28bbbfd06acbe523396072963418990e9dd"} Nov 29 00:39:12 crc kubenswrapper[4931]: I1129 00:39:12.986578 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" event={"ID":"054b5e53-28bc-4222-9f59-ad0ebd2cf767","Type":"ContainerStarted","Data":"3d9f418daa3c20c3537e6118648a8ed8d2d8e874a51d61f24feb04fb7eb23b28"} Nov 29 00:39:13 crc kubenswrapper[4931]: I1129 00:39:13.006460 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" podStartSLOduration=3.000224034 podStartE2EDuration="4.006443132s" podCreationTimestamp="2025-11-29 00:39:09 +0000 UTC" firstStartedPulling="2025-11-29 00:39:11.052587484 +0000 UTC m=+2364.214480716" lastFinishedPulling="2025-11-29 00:39:12.058806582 +0000 UTC m=+2365.220699814" observedRunningTime="2025-11-29 00:39:13.00498135 +0000 UTC m=+2366.166874612" watchObservedRunningTime="2025-11-29 00:39:13.006443132 +0000 UTC m=+2366.168336374" Nov 29 00:39:23 crc kubenswrapper[4931]: I1129 00:39:23.734869 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:39:23 crc kubenswrapper[4931]: I1129 00:39:23.735303 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:39:53 crc kubenswrapper[4931]: I1129 00:39:53.735422 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:39:53 crc kubenswrapper[4931]: I1129 00:39:53.736065 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:39:53 crc kubenswrapper[4931]: I1129 00:39:53.736113 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:39:53 crc kubenswrapper[4931]: I1129 00:39:53.736977 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 00:39:53 crc kubenswrapper[4931]: I1129 00:39:53.737047 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" gracePeriod=600 Nov 29 00:39:54 crc kubenswrapper[4931]: E1129 00:39:54.384758 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:39:54 crc kubenswrapper[4931]: I1129 00:39:54.442123 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" exitCode=0 Nov 29 00:39:54 crc kubenswrapper[4931]: I1129 00:39:54.442170 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2"} Nov 29 00:39:54 crc kubenswrapper[4931]: I1129 00:39:54.442203 4931 scope.go:117] "RemoveContainer" containerID="bc12cbef83d264cf45b46aedbbb0832fdfe3ed670622f291973be8df8b6edd8a" Nov 29 00:39:54 crc kubenswrapper[4931]: I1129 00:39:54.442942 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:39:54 crc kubenswrapper[4931]: E1129 00:39:54.443264 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:40:07 crc kubenswrapper[4931]: I1129 00:40:07.594689 4931 generic.go:334] "Generic (PLEG): container finished" podID="054b5e53-28bc-4222-9f59-ad0ebd2cf767" containerID="3d9f418daa3c20c3537e6118648a8ed8d2d8e874a51d61f24feb04fb7eb23b28" exitCode=0 Nov 29 00:40:07 crc kubenswrapper[4931]: I1129 00:40:07.594749 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" event={"ID":"054b5e53-28bc-4222-9f59-ad0ebd2cf767","Type":"ContainerDied","Data":"3d9f418daa3c20c3537e6118648a8ed8d2d8e874a51d61f24feb04fb7eb23b28"} Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.095525 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.212378 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-neutron-ovn-metadata-agent-neutron-config-0\") pod \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.212508 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-neutron-metadata-combined-ca-bundle\") pod \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.212575 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-inventory\") pod \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.212637 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-nova-metadata-neutron-config-0\") pod \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.212674 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-ssh-key\") pod \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.212758 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42lnk\" (UniqueName: \"kubernetes.io/projected/054b5e53-28bc-4222-9f59-ad0ebd2cf767-kube-api-access-42lnk\") pod \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\" (UID: \"054b5e53-28bc-4222-9f59-ad0ebd2cf767\") " Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.213348 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:40:09 crc kubenswrapper[4931]: E1129 00:40:09.213605 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.229127 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/054b5e53-28bc-4222-9f59-ad0ebd2cf767-kube-api-access-42lnk" (OuterVolumeSpecName: "kube-api-access-42lnk") pod "054b5e53-28bc-4222-9f59-ad0ebd2cf767" (UID: "054b5e53-28bc-4222-9f59-ad0ebd2cf767"). InnerVolumeSpecName "kube-api-access-42lnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.231119 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "054b5e53-28bc-4222-9f59-ad0ebd2cf767" (UID: "054b5e53-28bc-4222-9f59-ad0ebd2cf767"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.240544 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "054b5e53-28bc-4222-9f59-ad0ebd2cf767" (UID: "054b5e53-28bc-4222-9f59-ad0ebd2cf767"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.250890 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-inventory" (OuterVolumeSpecName: "inventory") pod "054b5e53-28bc-4222-9f59-ad0ebd2cf767" (UID: "054b5e53-28bc-4222-9f59-ad0ebd2cf767"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.262408 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "054b5e53-28bc-4222-9f59-ad0ebd2cf767" (UID: "054b5e53-28bc-4222-9f59-ad0ebd2cf767"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.266328 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "054b5e53-28bc-4222-9f59-ad0ebd2cf767" (UID: "054b5e53-28bc-4222-9f59-ad0ebd2cf767"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.316751 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.316802 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42lnk\" (UniqueName: \"kubernetes.io/projected/054b5e53-28bc-4222-9f59-ad0ebd2cf767-kube-api-access-42lnk\") on node \"crc\" DevicePath \"\"" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.316852 4931 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.316872 4931 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.316893 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.316911 4931 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/054b5e53-28bc-4222-9f59-ad0ebd2cf767-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.618969 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" event={"ID":"054b5e53-28bc-4222-9f59-ad0ebd2cf767","Type":"ContainerDied","Data":"b2becb29990a8278a6fe5aacf594b28bbbfd06acbe523396072963418990e9dd"} Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.619009 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2becb29990a8278a6fe5aacf594b28bbbfd06acbe523396072963418990e9dd" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.619049 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.743040 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958"] Nov 29 00:40:09 crc kubenswrapper[4931]: E1129 00:40:09.743401 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="054b5e53-28bc-4222-9f59-ad0ebd2cf767" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.743418 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="054b5e53-28bc-4222-9f59-ad0ebd2cf767" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.743615 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="054b5e53-28bc-4222-9f59-ad0ebd2cf767" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.744287 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.748468 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.748796 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.749114 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.749337 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.749487 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.758743 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958"] Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.929985 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.930201 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.930328 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7msd\" (UniqueName: \"kubernetes.io/projected/ca54c57b-1fac-46bf-aa71-830b5417fca0-kube-api-access-b7msd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.930682 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:09 crc kubenswrapper[4931]: I1129 00:40:09.930888 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.032838 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.032923 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.033007 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.033068 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.033099 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7msd\" (UniqueName: \"kubernetes.io/projected/ca54c57b-1fac-46bf-aa71-830b5417fca0-kube-api-access-b7msd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.039436 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.040039 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.040081 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.042153 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.064763 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7msd\" (UniqueName: \"kubernetes.io/projected/ca54c57b-1fac-46bf-aa71-830b5417fca0-kube-api-access-b7msd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-r5958\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.076238 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.479594 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958"] Nov 29 00:40:10 crc kubenswrapper[4931]: I1129 00:40:10.630570 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" event={"ID":"ca54c57b-1fac-46bf-aa71-830b5417fca0","Type":"ContainerStarted","Data":"f9185421eaa729640a873f30ce9586a5b66d4646f4c60e538e329d75e66a650a"} Nov 29 00:40:12 crc kubenswrapper[4931]: I1129 00:40:12.663588 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" event={"ID":"ca54c57b-1fac-46bf-aa71-830b5417fca0","Type":"ContainerStarted","Data":"e5ab68816ca3264f70f80cb336dee18f2e8111f8123fcaeeadc46d9a09a0ff2e"} Nov 29 00:40:12 crc kubenswrapper[4931]: I1129 00:40:12.681111 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" podStartSLOduration=1.892080096 podStartE2EDuration="3.681088248s" podCreationTimestamp="2025-11-29 00:40:09 +0000 UTC" firstStartedPulling="2025-11-29 00:40:10.479358087 +0000 UTC m=+2423.641251319" lastFinishedPulling="2025-11-29 00:40:12.268366199 +0000 UTC m=+2425.430259471" observedRunningTime="2025-11-29 00:40:12.677198987 +0000 UTC m=+2425.839092229" watchObservedRunningTime="2025-11-29 00:40:12.681088248 +0000 UTC m=+2425.842981490" Nov 29 00:40:16 crc kubenswrapper[4931]: I1129 00:40:16.446319 4931 scope.go:117] "RemoveContainer" containerID="32a3072cb42d291c32130cf2af084301795875b12d28133e5a63e517c62710bb" Nov 29 00:40:16 crc kubenswrapper[4931]: I1129 00:40:16.486408 4931 scope.go:117] "RemoveContainer" containerID="c63129adc265cddfe084643a7890b200411511049db208402c9f9958e322e5bf" Nov 29 00:40:16 crc kubenswrapper[4931]: I1129 00:40:16.563845 4931 scope.go:117] "RemoveContainer" containerID="18c9275c0835d2490eda45591ab92fdbf9451e3244532cd799177a505d3db268" Nov 29 00:40:20 crc kubenswrapper[4931]: I1129 00:40:20.213120 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:40:20 crc kubenswrapper[4931]: E1129 00:40:20.214482 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:40:35 crc kubenswrapper[4931]: I1129 00:40:35.212284 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:40:35 crc kubenswrapper[4931]: E1129 00:40:35.213389 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:40:47 crc kubenswrapper[4931]: I1129 00:40:47.223613 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:40:47 crc kubenswrapper[4931]: E1129 00:40:47.224647 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:41:02 crc kubenswrapper[4931]: I1129 00:41:02.211875 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:41:02 crc kubenswrapper[4931]: E1129 00:41:02.212652 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:41:15 crc kubenswrapper[4931]: I1129 00:41:15.212949 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:41:15 crc kubenswrapper[4931]: E1129 00:41:15.213855 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:41:28 crc kubenswrapper[4931]: I1129 00:41:28.212565 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:41:28 crc kubenswrapper[4931]: E1129 00:41:28.213343 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:41:42 crc kubenswrapper[4931]: I1129 00:41:42.212298 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:41:42 crc kubenswrapper[4931]: E1129 00:41:42.213493 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:41:57 crc kubenswrapper[4931]: I1129 00:41:57.228951 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:41:57 crc kubenswrapper[4931]: E1129 00:41:57.231080 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:42:12 crc kubenswrapper[4931]: I1129 00:42:12.213862 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:42:12 crc kubenswrapper[4931]: E1129 00:42:12.215263 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:42:23 crc kubenswrapper[4931]: I1129 00:42:23.212655 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:42:23 crc kubenswrapper[4931]: E1129 00:42:23.213901 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:42:37 crc kubenswrapper[4931]: I1129 00:42:37.225050 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:42:37 crc kubenswrapper[4931]: E1129 00:42:37.226113 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:42:51 crc kubenswrapper[4931]: I1129 00:42:51.212412 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:42:51 crc kubenswrapper[4931]: E1129 00:42:51.213555 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:43:05 crc kubenswrapper[4931]: I1129 00:43:05.213294 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:43:05 crc kubenswrapper[4931]: E1129 00:43:05.214714 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:43:16 crc kubenswrapper[4931]: I1129 00:43:16.211993 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:43:16 crc kubenswrapper[4931]: E1129 00:43:16.212889 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:43:30 crc kubenswrapper[4931]: I1129 00:43:30.214495 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:43:30 crc kubenswrapper[4931]: E1129 00:43:30.216160 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:43:42 crc kubenswrapper[4931]: I1129 00:43:42.212999 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:43:42 crc kubenswrapper[4931]: E1129 00:43:42.213934 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:43:56 crc kubenswrapper[4931]: I1129 00:43:56.212751 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:43:56 crc kubenswrapper[4931]: E1129 00:43:56.213740 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:44:08 crc kubenswrapper[4931]: I1129 00:44:08.213066 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:44:08 crc kubenswrapper[4931]: E1129 00:44:08.215482 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:44:20 crc kubenswrapper[4931]: I1129 00:44:20.212675 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:44:20 crc kubenswrapper[4931]: E1129 00:44:20.213710 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:44:34 crc kubenswrapper[4931]: I1129 00:44:34.212996 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:44:34 crc kubenswrapper[4931]: E1129 00:44:34.215195 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:44:45 crc kubenswrapper[4931]: I1129 00:44:45.224399 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:44:45 crc kubenswrapper[4931]: E1129 00:44:45.225983 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:44:58 crc kubenswrapper[4931]: I1129 00:44:58.212312 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:44:58 crc kubenswrapper[4931]: I1129 00:44:58.951682 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"3d459f7554777705e26d66be3e626f25d6b92931660a2f3ad9dff8fdaba9654f"} Nov 29 00:44:59 crc kubenswrapper[4931]: I1129 00:44:59.965030 4931 generic.go:334] "Generic (PLEG): container finished" podID="ca54c57b-1fac-46bf-aa71-830b5417fca0" containerID="e5ab68816ca3264f70f80cb336dee18f2e8111f8123fcaeeadc46d9a09a0ff2e" exitCode=0 Nov 29 00:44:59 crc kubenswrapper[4931]: I1129 00:44:59.965173 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" event={"ID":"ca54c57b-1fac-46bf-aa71-830b5417fca0","Type":"ContainerDied","Data":"e5ab68816ca3264f70f80cb336dee18f2e8111f8123fcaeeadc46d9a09a0ff2e"} Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.143362 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln"] Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.144903 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.147296 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.148002 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.160696 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln"] Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.220856 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk4r4\" (UniqueName: \"kubernetes.io/projected/d7148f32-d40a-4366-b070-29dcea134ee4-kube-api-access-hk4r4\") pod \"collect-profiles-29406285-6jvln\" (UID: \"d7148f32-d40a-4366-b070-29dcea134ee4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.220984 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7148f32-d40a-4366-b070-29dcea134ee4-config-volume\") pod \"collect-profiles-29406285-6jvln\" (UID: \"d7148f32-d40a-4366-b070-29dcea134ee4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.221051 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7148f32-d40a-4366-b070-29dcea134ee4-secret-volume\") pod \"collect-profiles-29406285-6jvln\" (UID: \"d7148f32-d40a-4366-b070-29dcea134ee4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.322660 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk4r4\" (UniqueName: \"kubernetes.io/projected/d7148f32-d40a-4366-b070-29dcea134ee4-kube-api-access-hk4r4\") pod \"collect-profiles-29406285-6jvln\" (UID: \"d7148f32-d40a-4366-b070-29dcea134ee4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.322985 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7148f32-d40a-4366-b070-29dcea134ee4-config-volume\") pod \"collect-profiles-29406285-6jvln\" (UID: \"d7148f32-d40a-4366-b070-29dcea134ee4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.323024 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7148f32-d40a-4366-b070-29dcea134ee4-secret-volume\") pod \"collect-profiles-29406285-6jvln\" (UID: \"d7148f32-d40a-4366-b070-29dcea134ee4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.325907 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7148f32-d40a-4366-b070-29dcea134ee4-config-volume\") pod \"collect-profiles-29406285-6jvln\" (UID: \"d7148f32-d40a-4366-b070-29dcea134ee4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.337015 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7148f32-d40a-4366-b070-29dcea134ee4-secret-volume\") pod \"collect-profiles-29406285-6jvln\" (UID: \"d7148f32-d40a-4366-b070-29dcea134ee4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.343320 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk4r4\" (UniqueName: \"kubernetes.io/projected/d7148f32-d40a-4366-b070-29dcea134ee4-kube-api-access-hk4r4\") pod \"collect-profiles-29406285-6jvln\" (UID: \"d7148f32-d40a-4366-b070-29dcea134ee4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.472909 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:00 crc kubenswrapper[4931]: I1129 00:45:00.963271 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln"] Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.301748 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.442463 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-libvirt-combined-ca-bundle\") pod \"ca54c57b-1fac-46bf-aa71-830b5417fca0\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.442563 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-inventory\") pod \"ca54c57b-1fac-46bf-aa71-830b5417fca0\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.442596 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-ssh-key\") pod \"ca54c57b-1fac-46bf-aa71-830b5417fca0\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.442647 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7msd\" (UniqueName: \"kubernetes.io/projected/ca54c57b-1fac-46bf-aa71-830b5417fca0-kube-api-access-b7msd\") pod \"ca54c57b-1fac-46bf-aa71-830b5417fca0\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.442694 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-libvirt-secret-0\") pod \"ca54c57b-1fac-46bf-aa71-830b5417fca0\" (UID: \"ca54c57b-1fac-46bf-aa71-830b5417fca0\") " Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.448288 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "ca54c57b-1fac-46bf-aa71-830b5417fca0" (UID: "ca54c57b-1fac-46bf-aa71-830b5417fca0"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.448978 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca54c57b-1fac-46bf-aa71-830b5417fca0-kube-api-access-b7msd" (OuterVolumeSpecName: "kube-api-access-b7msd") pod "ca54c57b-1fac-46bf-aa71-830b5417fca0" (UID: "ca54c57b-1fac-46bf-aa71-830b5417fca0"). InnerVolumeSpecName "kube-api-access-b7msd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.474243 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "ca54c57b-1fac-46bf-aa71-830b5417fca0" (UID: "ca54c57b-1fac-46bf-aa71-830b5417fca0"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.477948 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ca54c57b-1fac-46bf-aa71-830b5417fca0" (UID: "ca54c57b-1fac-46bf-aa71-830b5417fca0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.485512 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-inventory" (OuterVolumeSpecName: "inventory") pod "ca54c57b-1fac-46bf-aa71-830b5417fca0" (UID: "ca54c57b-1fac-46bf-aa71-830b5417fca0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.545367 4931 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.545402 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.545416 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.545428 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7msd\" (UniqueName: \"kubernetes.io/projected/ca54c57b-1fac-46bf-aa71-830b5417fca0-kube-api-access-b7msd\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.545441 4931 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ca54c57b-1fac-46bf-aa71-830b5417fca0-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.988787 4931 generic.go:334] "Generic (PLEG): container finished" podID="d7148f32-d40a-4366-b070-29dcea134ee4" containerID="a5f6230201dc056d3e95a92becc754d947f19989980f7105a4fda27580a3d34a" exitCode=0 Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.988881 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" event={"ID":"d7148f32-d40a-4366-b070-29dcea134ee4","Type":"ContainerDied","Data":"a5f6230201dc056d3e95a92becc754d947f19989980f7105a4fda27580a3d34a"} Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.989120 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" event={"ID":"d7148f32-d40a-4366-b070-29dcea134ee4","Type":"ContainerStarted","Data":"a012809ebe7d282d217867629b9b68a8efb033e7a83bd024638caa734dbb54a9"} Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.992152 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" event={"ID":"ca54c57b-1fac-46bf-aa71-830b5417fca0","Type":"ContainerDied","Data":"f9185421eaa729640a873f30ce9586a5b66d4646f4c60e538e329d75e66a650a"} Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.992191 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9185421eaa729640a873f30ce9586a5b66d4646f4c60e538e329d75e66a650a" Nov 29 00:45:01 crc kubenswrapper[4931]: I1129 00:45:01.992208 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-r5958" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.094327 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9"] Nov 29 00:45:02 crc kubenswrapper[4931]: E1129 00:45:02.095185 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca54c57b-1fac-46bf-aa71-830b5417fca0" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.095207 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca54c57b-1fac-46bf-aa71-830b5417fca0" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.095553 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca54c57b-1fac-46bf-aa71-830b5417fca0" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.096438 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.100121 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.100225 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.100398 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.100479 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.104173 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.104358 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.105156 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.109574 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9"] Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.156785 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.156847 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.156914 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.156934 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.157001 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.157065 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zdkh\" (UniqueName: \"kubernetes.io/projected/aa634449-4a1f-4827-a318-5502c395886e-kube-api-access-4zdkh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.157095 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.157113 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/aa634449-4a1f-4827-a318-5502c395886e-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.157215 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.259411 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.259466 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/aa634449-4a1f-4827-a318-5502c395886e-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.259587 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.259706 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.259747 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.259850 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.259889 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.259918 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.259979 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zdkh\" (UniqueName: \"kubernetes.io/projected/aa634449-4a1f-4827-a318-5502c395886e-kube-api-access-4zdkh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.261305 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/aa634449-4a1f-4827-a318-5502c395886e-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.267207 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.269867 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.270161 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.270849 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.271705 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.272529 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.272569 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.283541 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zdkh\" (UniqueName: \"kubernetes.io/projected/aa634449-4a1f-4827-a318-5502c395886e-kube-api-access-4zdkh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-srlj9\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:02 crc kubenswrapper[4931]: I1129 00:45:02.430025 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:45:03 crc kubenswrapper[4931]: I1129 00:45:03.080203 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9"] Nov 29 00:45:03 crc kubenswrapper[4931]: W1129 00:45:03.090661 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa634449_4a1f_4827_a318_5502c395886e.slice/crio-111e533d71e4d7fc4280406affb7a943bed1705c394cb08adc40fcecfb8ef51c WatchSource:0}: Error finding container 111e533d71e4d7fc4280406affb7a943bed1705c394cb08adc40fcecfb8ef51c: Status 404 returned error can't find the container with id 111e533d71e4d7fc4280406affb7a943bed1705c394cb08adc40fcecfb8ef51c Nov 29 00:45:03 crc kubenswrapper[4931]: I1129 00:45:03.094093 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 00:45:03 crc kubenswrapper[4931]: I1129 00:45:03.396264 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:03 crc kubenswrapper[4931]: I1129 00:45:03.490200 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7148f32-d40a-4366-b070-29dcea134ee4-config-volume\") pod \"d7148f32-d40a-4366-b070-29dcea134ee4\" (UID: \"d7148f32-d40a-4366-b070-29dcea134ee4\") " Nov 29 00:45:03 crc kubenswrapper[4931]: I1129 00:45:03.490356 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7148f32-d40a-4366-b070-29dcea134ee4-secret-volume\") pod \"d7148f32-d40a-4366-b070-29dcea134ee4\" (UID: \"d7148f32-d40a-4366-b070-29dcea134ee4\") " Nov 29 00:45:03 crc kubenswrapper[4931]: I1129 00:45:03.490386 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk4r4\" (UniqueName: \"kubernetes.io/projected/d7148f32-d40a-4366-b070-29dcea134ee4-kube-api-access-hk4r4\") pod \"d7148f32-d40a-4366-b070-29dcea134ee4\" (UID: \"d7148f32-d40a-4366-b070-29dcea134ee4\") " Nov 29 00:45:03 crc kubenswrapper[4931]: I1129 00:45:03.491377 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7148f32-d40a-4366-b070-29dcea134ee4-config-volume" (OuterVolumeSpecName: "config-volume") pod "d7148f32-d40a-4366-b070-29dcea134ee4" (UID: "d7148f32-d40a-4366-b070-29dcea134ee4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:45:03 crc kubenswrapper[4931]: I1129 00:45:03.499405 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7148f32-d40a-4366-b070-29dcea134ee4-kube-api-access-hk4r4" (OuterVolumeSpecName: "kube-api-access-hk4r4") pod "d7148f32-d40a-4366-b070-29dcea134ee4" (UID: "d7148f32-d40a-4366-b070-29dcea134ee4"). InnerVolumeSpecName "kube-api-access-hk4r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:45:03 crc kubenswrapper[4931]: I1129 00:45:03.499482 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7148f32-d40a-4366-b070-29dcea134ee4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d7148f32-d40a-4366-b070-29dcea134ee4" (UID: "d7148f32-d40a-4366-b070-29dcea134ee4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:45:03 crc kubenswrapper[4931]: I1129 00:45:03.593039 4931 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7148f32-d40a-4366-b070-29dcea134ee4-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:03 crc kubenswrapper[4931]: I1129 00:45:03.593096 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk4r4\" (UniqueName: \"kubernetes.io/projected/d7148f32-d40a-4366-b070-29dcea134ee4-kube-api-access-hk4r4\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:03 crc kubenswrapper[4931]: I1129 00:45:03.593115 4931 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7148f32-d40a-4366-b070-29dcea134ee4-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:04 crc kubenswrapper[4931]: I1129 00:45:04.014544 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" event={"ID":"aa634449-4a1f-4827-a318-5502c395886e","Type":"ContainerStarted","Data":"c7f2808cb1f1209ad78e538a0f709709f9e8db98bc5b9ffce793cda7c896b71e"} Nov 29 00:45:04 crc kubenswrapper[4931]: I1129 00:45:04.014999 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" event={"ID":"aa634449-4a1f-4827-a318-5502c395886e","Type":"ContainerStarted","Data":"111e533d71e4d7fc4280406affb7a943bed1705c394cb08adc40fcecfb8ef51c"} Nov 29 00:45:04 crc kubenswrapper[4931]: I1129 00:45:04.018025 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" event={"ID":"d7148f32-d40a-4366-b070-29dcea134ee4","Type":"ContainerDied","Data":"a012809ebe7d282d217867629b9b68a8efb033e7a83bd024638caa734dbb54a9"} Nov 29 00:45:04 crc kubenswrapper[4931]: I1129 00:45:04.018090 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a012809ebe7d282d217867629b9b68a8efb033e7a83bd024638caa734dbb54a9" Nov 29 00:45:04 crc kubenswrapper[4931]: I1129 00:45:04.018151 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406285-6jvln" Nov 29 00:45:04 crc kubenswrapper[4931]: I1129 00:45:04.037617 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" podStartSLOduration=1.487359755 podStartE2EDuration="2.037595822s" podCreationTimestamp="2025-11-29 00:45:02 +0000 UTC" firstStartedPulling="2025-11-29 00:45:03.093852139 +0000 UTC m=+2716.255745381" lastFinishedPulling="2025-11-29 00:45:03.644088186 +0000 UTC m=+2716.805981448" observedRunningTime="2025-11-29 00:45:04.035901884 +0000 UTC m=+2717.197795136" watchObservedRunningTime="2025-11-29 00:45:04.037595822 +0000 UTC m=+2717.199489054" Nov 29 00:45:04 crc kubenswrapper[4931]: I1129 00:45:04.498056 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7"] Nov 29 00:45:04 crc kubenswrapper[4931]: I1129 00:45:04.507867 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406240-fjsn7"] Nov 29 00:45:05 crc kubenswrapper[4931]: I1129 00:45:05.232939 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96bbcdc9-9b57-409e-b5b9-79a99034208a" path="/var/lib/kubelet/pods/96bbcdc9-9b57-409e-b5b9-79a99034208a/volumes" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.342267 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x4dhb"] Nov 29 00:45:10 crc kubenswrapper[4931]: E1129 00:45:10.343496 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7148f32-d40a-4366-b070-29dcea134ee4" containerName="collect-profiles" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.343519 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7148f32-d40a-4366-b070-29dcea134ee4" containerName="collect-profiles" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.343920 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7148f32-d40a-4366-b070-29dcea134ee4" containerName="collect-profiles" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.346099 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.357119 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x4dhb"] Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.437213 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2544c\" (UniqueName: \"kubernetes.io/projected/92aad049-525f-43f5-80da-87aecd20844b-kube-api-access-2544c\") pod \"redhat-marketplace-x4dhb\" (UID: \"92aad049-525f-43f5-80da-87aecd20844b\") " pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.437314 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92aad049-525f-43f5-80da-87aecd20844b-catalog-content\") pod \"redhat-marketplace-x4dhb\" (UID: \"92aad049-525f-43f5-80da-87aecd20844b\") " pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.437457 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92aad049-525f-43f5-80da-87aecd20844b-utilities\") pod \"redhat-marketplace-x4dhb\" (UID: \"92aad049-525f-43f5-80da-87aecd20844b\") " pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.539212 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92aad049-525f-43f5-80da-87aecd20844b-utilities\") pod \"redhat-marketplace-x4dhb\" (UID: \"92aad049-525f-43f5-80da-87aecd20844b\") " pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.539358 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2544c\" (UniqueName: \"kubernetes.io/projected/92aad049-525f-43f5-80da-87aecd20844b-kube-api-access-2544c\") pod \"redhat-marketplace-x4dhb\" (UID: \"92aad049-525f-43f5-80da-87aecd20844b\") " pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.539398 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92aad049-525f-43f5-80da-87aecd20844b-catalog-content\") pod \"redhat-marketplace-x4dhb\" (UID: \"92aad049-525f-43f5-80da-87aecd20844b\") " pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.539969 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92aad049-525f-43f5-80da-87aecd20844b-utilities\") pod \"redhat-marketplace-x4dhb\" (UID: \"92aad049-525f-43f5-80da-87aecd20844b\") " pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.539983 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92aad049-525f-43f5-80da-87aecd20844b-catalog-content\") pod \"redhat-marketplace-x4dhb\" (UID: \"92aad049-525f-43f5-80da-87aecd20844b\") " pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.559636 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2544c\" (UniqueName: \"kubernetes.io/projected/92aad049-525f-43f5-80da-87aecd20844b-kube-api-access-2544c\") pod \"redhat-marketplace-x4dhb\" (UID: \"92aad049-525f-43f5-80da-87aecd20844b\") " pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:10 crc kubenswrapper[4931]: I1129 00:45:10.674416 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:11 crc kubenswrapper[4931]: I1129 00:45:11.170474 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x4dhb"] Nov 29 00:45:11 crc kubenswrapper[4931]: W1129 00:45:11.176110 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92aad049_525f_43f5_80da_87aecd20844b.slice/crio-f0f92fb2665933c4c6bf6a0d8da4eaa7af2acb9b5c0e3ccb84e854c3da373d78 WatchSource:0}: Error finding container f0f92fb2665933c4c6bf6a0d8da4eaa7af2acb9b5c0e3ccb84e854c3da373d78: Status 404 returned error can't find the container with id f0f92fb2665933c4c6bf6a0d8da4eaa7af2acb9b5c0e3ccb84e854c3da373d78 Nov 29 00:45:12 crc kubenswrapper[4931]: I1129 00:45:12.116082 4931 generic.go:334] "Generic (PLEG): container finished" podID="92aad049-525f-43f5-80da-87aecd20844b" containerID="027ecfbd58a6c91b99fda4b79e951059ebec5bef9a96b892de5d5cc4ddbde285" exitCode=0 Nov 29 00:45:12 crc kubenswrapper[4931]: I1129 00:45:12.116202 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x4dhb" event={"ID":"92aad049-525f-43f5-80da-87aecd20844b","Type":"ContainerDied","Data":"027ecfbd58a6c91b99fda4b79e951059ebec5bef9a96b892de5d5cc4ddbde285"} Nov 29 00:45:12 crc kubenswrapper[4931]: I1129 00:45:12.116532 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x4dhb" event={"ID":"92aad049-525f-43f5-80da-87aecd20844b","Type":"ContainerStarted","Data":"f0f92fb2665933c4c6bf6a0d8da4eaa7af2acb9b5c0e3ccb84e854c3da373d78"} Nov 29 00:45:14 crc kubenswrapper[4931]: I1129 00:45:14.138642 4931 generic.go:334] "Generic (PLEG): container finished" podID="92aad049-525f-43f5-80da-87aecd20844b" containerID="ce28d78be7cce062bd097f6c7caebe27d1faee18b939f8c1020c6dab3fe6c5ae" exitCode=0 Nov 29 00:45:14 crc kubenswrapper[4931]: I1129 00:45:14.138751 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x4dhb" event={"ID":"92aad049-525f-43f5-80da-87aecd20844b","Type":"ContainerDied","Data":"ce28d78be7cce062bd097f6c7caebe27d1faee18b939f8c1020c6dab3fe6c5ae"} Nov 29 00:45:15 crc kubenswrapper[4931]: I1129 00:45:15.152881 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x4dhb" event={"ID":"92aad049-525f-43f5-80da-87aecd20844b","Type":"ContainerStarted","Data":"242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd"} Nov 29 00:45:15 crc kubenswrapper[4931]: I1129 00:45:15.182405 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x4dhb" podStartSLOduration=2.640500052 podStartE2EDuration="5.18237963s" podCreationTimestamp="2025-11-29 00:45:10 +0000 UTC" firstStartedPulling="2025-11-29 00:45:12.118672834 +0000 UTC m=+2725.280566076" lastFinishedPulling="2025-11-29 00:45:14.660552422 +0000 UTC m=+2727.822445654" observedRunningTime="2025-11-29 00:45:15.171301724 +0000 UTC m=+2728.333194996" watchObservedRunningTime="2025-11-29 00:45:15.18237963 +0000 UTC m=+2728.344272902" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.115667 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2ch9k"] Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.117780 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.158221 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2ch9k"] Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.251096 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f776c6cd-fc7d-4aca-b1dc-393ab3337964-catalog-content\") pod \"certified-operators-2ch9k\" (UID: \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\") " pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.251177 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f776c6cd-fc7d-4aca-b1dc-393ab3337964-utilities\") pod \"certified-operators-2ch9k\" (UID: \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\") " pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.251225 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmd45\" (UniqueName: \"kubernetes.io/projected/f776c6cd-fc7d-4aca-b1dc-393ab3337964-kube-api-access-xmd45\") pod \"certified-operators-2ch9k\" (UID: \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\") " pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.352719 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmd45\" (UniqueName: \"kubernetes.io/projected/f776c6cd-fc7d-4aca-b1dc-393ab3337964-kube-api-access-xmd45\") pod \"certified-operators-2ch9k\" (UID: \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\") " pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.352958 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f776c6cd-fc7d-4aca-b1dc-393ab3337964-catalog-content\") pod \"certified-operators-2ch9k\" (UID: \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\") " pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.353087 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f776c6cd-fc7d-4aca-b1dc-393ab3337964-utilities\") pod \"certified-operators-2ch9k\" (UID: \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\") " pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.353780 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f776c6cd-fc7d-4aca-b1dc-393ab3337964-utilities\") pod \"certified-operators-2ch9k\" (UID: \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\") " pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.353782 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f776c6cd-fc7d-4aca-b1dc-393ab3337964-catalog-content\") pod \"certified-operators-2ch9k\" (UID: \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\") " pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.374056 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmd45\" (UniqueName: \"kubernetes.io/projected/f776c6cd-fc7d-4aca-b1dc-393ab3337964-kube-api-access-xmd45\") pod \"certified-operators-2ch9k\" (UID: \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\") " pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.437318 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.757040 4931 scope.go:117] "RemoveContainer" containerID="a0faded80efe8587290a77ed570ea80cd569c557fbf016a8c3089016d33ab90b" Nov 29 00:45:16 crc kubenswrapper[4931]: I1129 00:45:16.961290 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2ch9k"] Nov 29 00:45:17 crc kubenswrapper[4931]: I1129 00:45:17.170692 4931 generic.go:334] "Generic (PLEG): container finished" podID="f776c6cd-fc7d-4aca-b1dc-393ab3337964" containerID="fc8f67c634b6a80590ef7eb6eabe0542fd20545c747e4f05c4991294dd8acce3" exitCode=0 Nov 29 00:45:17 crc kubenswrapper[4931]: I1129 00:45:17.170765 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2ch9k" event={"ID":"f776c6cd-fc7d-4aca-b1dc-393ab3337964","Type":"ContainerDied","Data":"fc8f67c634b6a80590ef7eb6eabe0542fd20545c747e4f05c4991294dd8acce3"} Nov 29 00:45:17 crc kubenswrapper[4931]: I1129 00:45:17.170797 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2ch9k" event={"ID":"f776c6cd-fc7d-4aca-b1dc-393ab3337964","Type":"ContainerStarted","Data":"6e8c4832846c92813dcbaa51e5f5fb012e0ec710f5fe76becd42847477f7c39b"} Nov 29 00:45:18 crc kubenswrapper[4931]: I1129 00:45:18.186569 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2ch9k" event={"ID":"f776c6cd-fc7d-4aca-b1dc-393ab3337964","Type":"ContainerStarted","Data":"b6ea0dbdfd892c3c7fdc7d618fd1d2e91889935425581059f570cf35591d1104"} Nov 29 00:45:19 crc kubenswrapper[4931]: I1129 00:45:19.195295 4931 generic.go:334] "Generic (PLEG): container finished" podID="f776c6cd-fc7d-4aca-b1dc-393ab3337964" containerID="b6ea0dbdfd892c3c7fdc7d618fd1d2e91889935425581059f570cf35591d1104" exitCode=0 Nov 29 00:45:19 crc kubenswrapper[4931]: I1129 00:45:19.195346 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2ch9k" event={"ID":"f776c6cd-fc7d-4aca-b1dc-393ab3337964","Type":"ContainerDied","Data":"b6ea0dbdfd892c3c7fdc7d618fd1d2e91889935425581059f570cf35591d1104"} Nov 29 00:45:20 crc kubenswrapper[4931]: I1129 00:45:20.209145 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2ch9k" event={"ID":"f776c6cd-fc7d-4aca-b1dc-393ab3337964","Type":"ContainerStarted","Data":"dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873"} Nov 29 00:45:20 crc kubenswrapper[4931]: I1129 00:45:20.245909 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2ch9k" podStartSLOduration=1.749376187 podStartE2EDuration="4.245889678s" podCreationTimestamp="2025-11-29 00:45:16 +0000 UTC" firstStartedPulling="2025-11-29 00:45:17.172947436 +0000 UTC m=+2730.334840698" lastFinishedPulling="2025-11-29 00:45:19.669460917 +0000 UTC m=+2732.831354189" observedRunningTime="2025-11-29 00:45:20.228798179 +0000 UTC m=+2733.390691491" watchObservedRunningTime="2025-11-29 00:45:20.245889678 +0000 UTC m=+2733.407782910" Nov 29 00:45:20 crc kubenswrapper[4931]: I1129 00:45:20.675060 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:20 crc kubenswrapper[4931]: I1129 00:45:20.675451 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:20 crc kubenswrapper[4931]: I1129 00:45:20.744661 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:21 crc kubenswrapper[4931]: I1129 00:45:21.276330 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:22 crc kubenswrapper[4931]: I1129 00:45:22.916869 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x4dhb"] Nov 29 00:45:23 crc kubenswrapper[4931]: I1129 00:45:23.235741 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x4dhb" podUID="92aad049-525f-43f5-80da-87aecd20844b" containerName="registry-server" containerID="cri-o://242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd" gracePeriod=2 Nov 29 00:45:23 crc kubenswrapper[4931]: I1129 00:45:23.776038 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:23 crc kubenswrapper[4931]: I1129 00:45:23.803324 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92aad049-525f-43f5-80da-87aecd20844b-catalog-content\") pod \"92aad049-525f-43f5-80da-87aecd20844b\" (UID: \"92aad049-525f-43f5-80da-87aecd20844b\") " Nov 29 00:45:23 crc kubenswrapper[4931]: I1129 00:45:23.803379 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2544c\" (UniqueName: \"kubernetes.io/projected/92aad049-525f-43f5-80da-87aecd20844b-kube-api-access-2544c\") pod \"92aad049-525f-43f5-80da-87aecd20844b\" (UID: \"92aad049-525f-43f5-80da-87aecd20844b\") " Nov 29 00:45:23 crc kubenswrapper[4931]: I1129 00:45:23.803472 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92aad049-525f-43f5-80da-87aecd20844b-utilities\") pod \"92aad049-525f-43f5-80da-87aecd20844b\" (UID: \"92aad049-525f-43f5-80da-87aecd20844b\") " Nov 29 00:45:23 crc kubenswrapper[4931]: I1129 00:45:23.804538 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92aad049-525f-43f5-80da-87aecd20844b-utilities" (OuterVolumeSpecName: "utilities") pod "92aad049-525f-43f5-80da-87aecd20844b" (UID: "92aad049-525f-43f5-80da-87aecd20844b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:45:23 crc kubenswrapper[4931]: I1129 00:45:23.812689 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92aad049-525f-43f5-80da-87aecd20844b-kube-api-access-2544c" (OuterVolumeSpecName: "kube-api-access-2544c") pod "92aad049-525f-43f5-80da-87aecd20844b" (UID: "92aad049-525f-43f5-80da-87aecd20844b"). InnerVolumeSpecName "kube-api-access-2544c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:45:23 crc kubenswrapper[4931]: I1129 00:45:23.829302 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92aad049-525f-43f5-80da-87aecd20844b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92aad049-525f-43f5-80da-87aecd20844b" (UID: "92aad049-525f-43f5-80da-87aecd20844b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:45:23 crc kubenswrapper[4931]: I1129 00:45:23.905561 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92aad049-525f-43f5-80da-87aecd20844b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:23 crc kubenswrapper[4931]: I1129 00:45:23.905618 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2544c\" (UniqueName: \"kubernetes.io/projected/92aad049-525f-43f5-80da-87aecd20844b-kube-api-access-2544c\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:23 crc kubenswrapper[4931]: I1129 00:45:23.905640 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92aad049-525f-43f5-80da-87aecd20844b-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.251945 4931 generic.go:334] "Generic (PLEG): container finished" podID="92aad049-525f-43f5-80da-87aecd20844b" containerID="242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd" exitCode=0 Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.252001 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x4dhb" event={"ID":"92aad049-525f-43f5-80da-87aecd20844b","Type":"ContainerDied","Data":"242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd"} Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.252062 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x4dhb" event={"ID":"92aad049-525f-43f5-80da-87aecd20844b","Type":"ContainerDied","Data":"f0f92fb2665933c4c6bf6a0d8da4eaa7af2acb9b5c0e3ccb84e854c3da373d78"} Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.252069 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x4dhb" Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.252083 4931 scope.go:117] "RemoveContainer" containerID="242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd" Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.288228 4931 scope.go:117] "RemoveContainer" containerID="ce28d78be7cce062bd097f6c7caebe27d1faee18b939f8c1020c6dab3fe6c5ae" Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.304312 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x4dhb"] Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.318672 4931 scope.go:117] "RemoveContainer" containerID="027ecfbd58a6c91b99fda4b79e951059ebec5bef9a96b892de5d5cc4ddbde285" Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.327628 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x4dhb"] Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.378597 4931 scope.go:117] "RemoveContainer" containerID="242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd" Nov 29 00:45:24 crc kubenswrapper[4931]: E1129 00:45:24.379212 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd\": container with ID starting with 242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd not found: ID does not exist" containerID="242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd" Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.379290 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd"} err="failed to get container status \"242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd\": rpc error: code = NotFound desc = could not find container \"242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd\": container with ID starting with 242f33e4b3833d2f5e8b6a6bf5dc50f0da1cb203b15d1d8222f680754ed28edd not found: ID does not exist" Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.379325 4931 scope.go:117] "RemoveContainer" containerID="ce28d78be7cce062bd097f6c7caebe27d1faee18b939f8c1020c6dab3fe6c5ae" Nov 29 00:45:24 crc kubenswrapper[4931]: E1129 00:45:24.379983 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce28d78be7cce062bd097f6c7caebe27d1faee18b939f8c1020c6dab3fe6c5ae\": container with ID starting with ce28d78be7cce062bd097f6c7caebe27d1faee18b939f8c1020c6dab3fe6c5ae not found: ID does not exist" containerID="ce28d78be7cce062bd097f6c7caebe27d1faee18b939f8c1020c6dab3fe6c5ae" Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.380034 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce28d78be7cce062bd097f6c7caebe27d1faee18b939f8c1020c6dab3fe6c5ae"} err="failed to get container status \"ce28d78be7cce062bd097f6c7caebe27d1faee18b939f8c1020c6dab3fe6c5ae\": rpc error: code = NotFound desc = could not find container \"ce28d78be7cce062bd097f6c7caebe27d1faee18b939f8c1020c6dab3fe6c5ae\": container with ID starting with ce28d78be7cce062bd097f6c7caebe27d1faee18b939f8c1020c6dab3fe6c5ae not found: ID does not exist" Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.380068 4931 scope.go:117] "RemoveContainer" containerID="027ecfbd58a6c91b99fda4b79e951059ebec5bef9a96b892de5d5cc4ddbde285" Nov 29 00:45:24 crc kubenswrapper[4931]: E1129 00:45:24.380498 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"027ecfbd58a6c91b99fda4b79e951059ebec5bef9a96b892de5d5cc4ddbde285\": container with ID starting with 027ecfbd58a6c91b99fda4b79e951059ebec5bef9a96b892de5d5cc4ddbde285 not found: ID does not exist" containerID="027ecfbd58a6c91b99fda4b79e951059ebec5bef9a96b892de5d5cc4ddbde285" Nov 29 00:45:24 crc kubenswrapper[4931]: I1129 00:45:24.380525 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"027ecfbd58a6c91b99fda4b79e951059ebec5bef9a96b892de5d5cc4ddbde285"} err="failed to get container status \"027ecfbd58a6c91b99fda4b79e951059ebec5bef9a96b892de5d5cc4ddbde285\": rpc error: code = NotFound desc = could not find container \"027ecfbd58a6c91b99fda4b79e951059ebec5bef9a96b892de5d5cc4ddbde285\": container with ID starting with 027ecfbd58a6c91b99fda4b79e951059ebec5bef9a96b892de5d5cc4ddbde285 not found: ID does not exist" Nov 29 00:45:25 crc kubenswrapper[4931]: I1129 00:45:25.224682 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92aad049-525f-43f5-80da-87aecd20844b" path="/var/lib/kubelet/pods/92aad049-525f-43f5-80da-87aecd20844b/volumes" Nov 29 00:45:26 crc kubenswrapper[4931]: I1129 00:45:26.437716 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:26 crc kubenswrapper[4931]: I1129 00:45:26.437946 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:26 crc kubenswrapper[4931]: I1129 00:45:26.516020 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:27 crc kubenswrapper[4931]: I1129 00:45:27.347081 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:28 crc kubenswrapper[4931]: I1129 00:45:28.110761 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2ch9k"] Nov 29 00:45:29 crc kubenswrapper[4931]: I1129 00:45:29.326515 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2ch9k" podUID="f776c6cd-fc7d-4aca-b1dc-393ab3337964" containerName="registry-server" containerID="cri-o://dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873" gracePeriod=2 Nov 29 00:45:29 crc kubenswrapper[4931]: I1129 00:45:29.917338 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.034954 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f776c6cd-fc7d-4aca-b1dc-393ab3337964-utilities\") pod \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\" (UID: \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\") " Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.035077 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmd45\" (UniqueName: \"kubernetes.io/projected/f776c6cd-fc7d-4aca-b1dc-393ab3337964-kube-api-access-xmd45\") pod \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\" (UID: \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\") " Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.035154 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f776c6cd-fc7d-4aca-b1dc-393ab3337964-catalog-content\") pod \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\" (UID: \"f776c6cd-fc7d-4aca-b1dc-393ab3337964\") " Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.036208 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f776c6cd-fc7d-4aca-b1dc-393ab3337964-utilities" (OuterVolumeSpecName: "utilities") pod "f776c6cd-fc7d-4aca-b1dc-393ab3337964" (UID: "f776c6cd-fc7d-4aca-b1dc-393ab3337964"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.044245 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f776c6cd-fc7d-4aca-b1dc-393ab3337964-kube-api-access-xmd45" (OuterVolumeSpecName: "kube-api-access-xmd45") pod "f776c6cd-fc7d-4aca-b1dc-393ab3337964" (UID: "f776c6cd-fc7d-4aca-b1dc-393ab3337964"). InnerVolumeSpecName "kube-api-access-xmd45". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.108545 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f776c6cd-fc7d-4aca-b1dc-393ab3337964-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f776c6cd-fc7d-4aca-b1dc-393ab3337964" (UID: "f776c6cd-fc7d-4aca-b1dc-393ab3337964"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.137305 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f776c6cd-fc7d-4aca-b1dc-393ab3337964-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.137336 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmd45\" (UniqueName: \"kubernetes.io/projected/f776c6cd-fc7d-4aca-b1dc-393ab3337964-kube-api-access-xmd45\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.137349 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f776c6cd-fc7d-4aca-b1dc-393ab3337964-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.340763 4931 generic.go:334] "Generic (PLEG): container finished" podID="f776c6cd-fc7d-4aca-b1dc-393ab3337964" containerID="dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873" exitCode=0 Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.340865 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2ch9k" event={"ID":"f776c6cd-fc7d-4aca-b1dc-393ab3337964","Type":"ContainerDied","Data":"dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873"} Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.340937 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2ch9k" event={"ID":"f776c6cd-fc7d-4aca-b1dc-393ab3337964","Type":"ContainerDied","Data":"6e8c4832846c92813dcbaa51e5f5fb012e0ec710f5fe76becd42847477f7c39b"} Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.340955 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2ch9k" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.340964 4931 scope.go:117] "RemoveContainer" containerID="dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.372854 4931 scope.go:117] "RemoveContainer" containerID="b6ea0dbdfd892c3c7fdc7d618fd1d2e91889935425581059f570cf35591d1104" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.401508 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2ch9k"] Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.413732 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2ch9k"] Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.432148 4931 scope.go:117] "RemoveContainer" containerID="fc8f67c634b6a80590ef7eb6eabe0542fd20545c747e4f05c4991294dd8acce3" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.473084 4931 scope.go:117] "RemoveContainer" containerID="dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873" Nov 29 00:45:30 crc kubenswrapper[4931]: E1129 00:45:30.473614 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873\": container with ID starting with dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873 not found: ID does not exist" containerID="dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.473667 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873"} err="failed to get container status \"dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873\": rpc error: code = NotFound desc = could not find container \"dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873\": container with ID starting with dbe971ca86995d87273b0629cc6bc5d0c83354b2b9f126ceda8c7060ac691873 not found: ID does not exist" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.473699 4931 scope.go:117] "RemoveContainer" containerID="b6ea0dbdfd892c3c7fdc7d618fd1d2e91889935425581059f570cf35591d1104" Nov 29 00:45:30 crc kubenswrapper[4931]: E1129 00:45:30.474295 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6ea0dbdfd892c3c7fdc7d618fd1d2e91889935425581059f570cf35591d1104\": container with ID starting with b6ea0dbdfd892c3c7fdc7d618fd1d2e91889935425581059f570cf35591d1104 not found: ID does not exist" containerID="b6ea0dbdfd892c3c7fdc7d618fd1d2e91889935425581059f570cf35591d1104" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.474347 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6ea0dbdfd892c3c7fdc7d618fd1d2e91889935425581059f570cf35591d1104"} err="failed to get container status \"b6ea0dbdfd892c3c7fdc7d618fd1d2e91889935425581059f570cf35591d1104\": rpc error: code = NotFound desc = could not find container \"b6ea0dbdfd892c3c7fdc7d618fd1d2e91889935425581059f570cf35591d1104\": container with ID starting with b6ea0dbdfd892c3c7fdc7d618fd1d2e91889935425581059f570cf35591d1104 not found: ID does not exist" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.474376 4931 scope.go:117] "RemoveContainer" containerID="fc8f67c634b6a80590ef7eb6eabe0542fd20545c747e4f05c4991294dd8acce3" Nov 29 00:45:30 crc kubenswrapper[4931]: E1129 00:45:30.474628 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc8f67c634b6a80590ef7eb6eabe0542fd20545c747e4f05c4991294dd8acce3\": container with ID starting with fc8f67c634b6a80590ef7eb6eabe0542fd20545c747e4f05c4991294dd8acce3 not found: ID does not exist" containerID="fc8f67c634b6a80590ef7eb6eabe0542fd20545c747e4f05c4991294dd8acce3" Nov 29 00:45:30 crc kubenswrapper[4931]: I1129 00:45:30.474652 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc8f67c634b6a80590ef7eb6eabe0542fd20545c747e4f05c4991294dd8acce3"} err="failed to get container status \"fc8f67c634b6a80590ef7eb6eabe0542fd20545c747e4f05c4991294dd8acce3\": rpc error: code = NotFound desc = could not find container \"fc8f67c634b6a80590ef7eb6eabe0542fd20545c747e4f05c4991294dd8acce3\": container with ID starting with fc8f67c634b6a80590ef7eb6eabe0542fd20545c747e4f05c4991294dd8acce3 not found: ID does not exist" Nov 29 00:45:31 crc kubenswrapper[4931]: I1129 00:45:31.228588 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f776c6cd-fc7d-4aca-b1dc-393ab3337964" path="/var/lib/kubelet/pods/f776c6cd-fc7d-4aca-b1dc-393ab3337964/volumes" Nov 29 00:47:23 crc kubenswrapper[4931]: I1129 00:47:23.735424 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:47:23 crc kubenswrapper[4931]: I1129 00:47:23.736086 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:47:53 crc kubenswrapper[4931]: I1129 00:47:53.735173 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:47:53 crc kubenswrapper[4931]: I1129 00:47:53.735947 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:48:17 crc kubenswrapper[4931]: I1129 00:48:17.181218 4931 generic.go:334] "Generic (PLEG): container finished" podID="aa634449-4a1f-4827-a318-5502c395886e" containerID="c7f2808cb1f1209ad78e538a0f709709f9e8db98bc5b9ffce793cda7c896b71e" exitCode=0 Nov 29 00:48:17 crc kubenswrapper[4931]: I1129 00:48:17.181359 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" event={"ID":"aa634449-4a1f-4827-a318-5502c395886e","Type":"ContainerDied","Data":"c7f2808cb1f1209ad78e538a0f709709f9e8db98bc5b9ffce793cda7c896b71e"} Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.626255 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.740405 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-combined-ca-bundle\") pod \"aa634449-4a1f-4827-a318-5502c395886e\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.740537 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-cell1-compute-config-0\") pod \"aa634449-4a1f-4827-a318-5502c395886e\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.740562 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-ssh-key\") pod \"aa634449-4a1f-4827-a318-5502c395886e\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.740640 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/aa634449-4a1f-4827-a318-5502c395886e-nova-extra-config-0\") pod \"aa634449-4a1f-4827-a318-5502c395886e\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.740680 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-migration-ssh-key-0\") pod \"aa634449-4a1f-4827-a318-5502c395886e\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.740730 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-migration-ssh-key-1\") pod \"aa634449-4a1f-4827-a318-5502c395886e\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.740765 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-cell1-compute-config-1\") pod \"aa634449-4a1f-4827-a318-5502c395886e\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.740865 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-inventory\") pod \"aa634449-4a1f-4827-a318-5502c395886e\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.740901 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zdkh\" (UniqueName: \"kubernetes.io/projected/aa634449-4a1f-4827-a318-5502c395886e-kube-api-access-4zdkh\") pod \"aa634449-4a1f-4827-a318-5502c395886e\" (UID: \"aa634449-4a1f-4827-a318-5502c395886e\") " Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.747990 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "aa634449-4a1f-4827-a318-5502c395886e" (UID: "aa634449-4a1f-4827-a318-5502c395886e"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.748507 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa634449-4a1f-4827-a318-5502c395886e-kube-api-access-4zdkh" (OuterVolumeSpecName: "kube-api-access-4zdkh") pod "aa634449-4a1f-4827-a318-5502c395886e" (UID: "aa634449-4a1f-4827-a318-5502c395886e"). InnerVolumeSpecName "kube-api-access-4zdkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.774934 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-inventory" (OuterVolumeSpecName: "inventory") pod "aa634449-4a1f-4827-a318-5502c395886e" (UID: "aa634449-4a1f-4827-a318-5502c395886e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.776609 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa634449-4a1f-4827-a318-5502c395886e-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "aa634449-4a1f-4827-a318-5502c395886e" (UID: "aa634449-4a1f-4827-a318-5502c395886e"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.783311 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aa634449-4a1f-4827-a318-5502c395886e" (UID: "aa634449-4a1f-4827-a318-5502c395886e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.792177 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "aa634449-4a1f-4827-a318-5502c395886e" (UID: "aa634449-4a1f-4827-a318-5502c395886e"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.792840 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "aa634449-4a1f-4827-a318-5502c395886e" (UID: "aa634449-4a1f-4827-a318-5502c395886e"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.795227 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "aa634449-4a1f-4827-a318-5502c395886e" (UID: "aa634449-4a1f-4827-a318-5502c395886e"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.803235 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "aa634449-4a1f-4827-a318-5502c395886e" (UID: "aa634449-4a1f-4827-a318-5502c395886e"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.843929 4931 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.843967 4931 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.843978 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.843987 4931 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/aa634449-4a1f-4827-a318-5502c395886e-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.843997 4931 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.844005 4931 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.844014 4931 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.844024 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa634449-4a1f-4827-a318-5502c395886e-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:48:18 crc kubenswrapper[4931]: I1129 00:48:18.844034 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zdkh\" (UniqueName: \"kubernetes.io/projected/aa634449-4a1f-4827-a318-5502c395886e-kube-api-access-4zdkh\") on node \"crc\" DevicePath \"\"" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.200105 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" event={"ID":"aa634449-4a1f-4827-a318-5502c395886e","Type":"ContainerDied","Data":"111e533d71e4d7fc4280406affb7a943bed1705c394cb08adc40fcecfb8ef51c"} Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.200149 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="111e533d71e4d7fc4280406affb7a943bed1705c394cb08adc40fcecfb8ef51c" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.200200 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-srlj9" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.333960 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj"] Nov 29 00:48:19 crc kubenswrapper[4931]: E1129 00:48:19.334399 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92aad049-525f-43f5-80da-87aecd20844b" containerName="registry-server" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.334418 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="92aad049-525f-43f5-80da-87aecd20844b" containerName="registry-server" Nov 29 00:48:19 crc kubenswrapper[4931]: E1129 00:48:19.334432 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92aad049-525f-43f5-80da-87aecd20844b" containerName="extract-utilities" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.334439 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="92aad049-525f-43f5-80da-87aecd20844b" containerName="extract-utilities" Nov 29 00:48:19 crc kubenswrapper[4931]: E1129 00:48:19.334452 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa634449-4a1f-4827-a318-5502c395886e" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.334458 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa634449-4a1f-4827-a318-5502c395886e" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 29 00:48:19 crc kubenswrapper[4931]: E1129 00:48:19.334480 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f776c6cd-fc7d-4aca-b1dc-393ab3337964" containerName="extract-content" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.334487 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f776c6cd-fc7d-4aca-b1dc-393ab3337964" containerName="extract-content" Nov 29 00:48:19 crc kubenswrapper[4931]: E1129 00:48:19.334498 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f776c6cd-fc7d-4aca-b1dc-393ab3337964" containerName="registry-server" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.334503 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f776c6cd-fc7d-4aca-b1dc-393ab3337964" containerName="registry-server" Nov 29 00:48:19 crc kubenswrapper[4931]: E1129 00:48:19.334514 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92aad049-525f-43f5-80da-87aecd20844b" containerName="extract-content" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.334519 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="92aad049-525f-43f5-80da-87aecd20844b" containerName="extract-content" Nov 29 00:48:19 crc kubenswrapper[4931]: E1129 00:48:19.334537 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f776c6cd-fc7d-4aca-b1dc-393ab3337964" containerName="extract-utilities" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.334544 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="f776c6cd-fc7d-4aca-b1dc-393ab3337964" containerName="extract-utilities" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.334707 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa634449-4a1f-4827-a318-5502c395886e" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.334724 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="92aad049-525f-43f5-80da-87aecd20844b" containerName="registry-server" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.334731 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="f776c6cd-fc7d-4aca-b1dc-393ab3337964" containerName="registry-server" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.335325 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.337765 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.337765 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kczjh" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.339740 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.339929 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.340002 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.344503 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj"] Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.453503 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.453926 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.453987 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.454186 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.454250 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkt66\" (UniqueName: \"kubernetes.io/projected/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-kube-api-access-tkt66\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.454341 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.454387 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.556016 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.556184 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.556237 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.556377 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.556487 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.556544 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkt66\" (UniqueName: \"kubernetes.io/projected/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-kube-api-access-tkt66\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.556623 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.561166 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.561320 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.561711 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.562182 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.563143 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.568729 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.580574 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkt66\" (UniqueName: \"kubernetes.io/projected/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-kube-api-access-tkt66\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:19 crc kubenswrapper[4931]: I1129 00:48:19.665977 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:48:20 crc kubenswrapper[4931]: I1129 00:48:20.238951 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj"] Nov 29 00:48:21 crc kubenswrapper[4931]: I1129 00:48:21.235083 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" event={"ID":"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a","Type":"ContainerStarted","Data":"7be66c7a2656298d93f7fdd522b9202e98c7d90d7983dab23b7502908b715bcb"} Nov 29 00:48:22 crc kubenswrapper[4931]: I1129 00:48:22.248971 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" event={"ID":"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a","Type":"ContainerStarted","Data":"07cea9ced249a9827f6811ad7c6901bd203497cc31b9dc8f5c00896a5789ee2b"} Nov 29 00:48:22 crc kubenswrapper[4931]: I1129 00:48:22.277273 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" podStartSLOduration=2.530590941 podStartE2EDuration="3.277246172s" podCreationTimestamp="2025-11-29 00:48:19 +0000 UTC" firstStartedPulling="2025-11-29 00:48:20.244088757 +0000 UTC m=+2913.405981999" lastFinishedPulling="2025-11-29 00:48:20.990743988 +0000 UTC m=+2914.152637230" observedRunningTime="2025-11-29 00:48:22.270934832 +0000 UTC m=+2915.432828094" watchObservedRunningTime="2025-11-29 00:48:22.277246172 +0000 UTC m=+2915.439139444" Nov 29 00:48:23 crc kubenswrapper[4931]: I1129 00:48:23.735223 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:48:23 crc kubenswrapper[4931]: I1129 00:48:23.735330 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:48:23 crc kubenswrapper[4931]: I1129 00:48:23.735410 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:48:23 crc kubenswrapper[4931]: I1129 00:48:23.736709 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3d459f7554777705e26d66be3e626f25d6b92931660a2f3ad9dff8fdaba9654f"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 00:48:23 crc kubenswrapper[4931]: I1129 00:48:23.736906 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://3d459f7554777705e26d66be3e626f25d6b92931660a2f3ad9dff8fdaba9654f" gracePeriod=600 Nov 29 00:48:24 crc kubenswrapper[4931]: I1129 00:48:24.274722 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="3d459f7554777705e26d66be3e626f25d6b92931660a2f3ad9dff8fdaba9654f" exitCode=0 Nov 29 00:48:24 crc kubenswrapper[4931]: I1129 00:48:24.275055 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"3d459f7554777705e26d66be3e626f25d6b92931660a2f3ad9dff8fdaba9654f"} Nov 29 00:48:24 crc kubenswrapper[4931]: I1129 00:48:24.275535 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251"} Nov 29 00:48:24 crc kubenswrapper[4931]: I1129 00:48:24.275568 4931 scope.go:117] "RemoveContainer" containerID="becbba4d3cdda387be1f695aeec89f4c5b0968e9deabdfdb7764055e6f65c8d2" Nov 29 00:48:47 crc kubenswrapper[4931]: I1129 00:48:47.672743 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p4zkb"] Nov 29 00:48:47 crc kubenswrapper[4931]: I1129 00:48:47.676712 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:47 crc kubenswrapper[4931]: I1129 00:48:47.694394 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4zkb"] Nov 29 00:48:47 crc kubenswrapper[4931]: I1129 00:48:47.705656 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c19ec362-e000-4dae-8b3c-f0d3863c96f5-utilities\") pod \"community-operators-p4zkb\" (UID: \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\") " pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:47 crc kubenswrapper[4931]: I1129 00:48:47.705761 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxv5p\" (UniqueName: \"kubernetes.io/projected/c19ec362-e000-4dae-8b3c-f0d3863c96f5-kube-api-access-fxv5p\") pod \"community-operators-p4zkb\" (UID: \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\") " pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:47 crc kubenswrapper[4931]: I1129 00:48:47.705878 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c19ec362-e000-4dae-8b3c-f0d3863c96f5-catalog-content\") pod \"community-operators-p4zkb\" (UID: \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\") " pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:47 crc kubenswrapper[4931]: I1129 00:48:47.807463 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c19ec362-e000-4dae-8b3c-f0d3863c96f5-utilities\") pod \"community-operators-p4zkb\" (UID: \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\") " pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:47 crc kubenswrapper[4931]: I1129 00:48:47.807576 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxv5p\" (UniqueName: \"kubernetes.io/projected/c19ec362-e000-4dae-8b3c-f0d3863c96f5-kube-api-access-fxv5p\") pod \"community-operators-p4zkb\" (UID: \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\") " pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:47 crc kubenswrapper[4931]: I1129 00:48:47.807659 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c19ec362-e000-4dae-8b3c-f0d3863c96f5-catalog-content\") pod \"community-operators-p4zkb\" (UID: \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\") " pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:47 crc kubenswrapper[4931]: I1129 00:48:47.808293 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c19ec362-e000-4dae-8b3c-f0d3863c96f5-catalog-content\") pod \"community-operators-p4zkb\" (UID: \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\") " pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:47 crc kubenswrapper[4931]: I1129 00:48:47.808324 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c19ec362-e000-4dae-8b3c-f0d3863c96f5-utilities\") pod \"community-operators-p4zkb\" (UID: \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\") " pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:47 crc kubenswrapper[4931]: I1129 00:48:47.834194 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxv5p\" (UniqueName: \"kubernetes.io/projected/c19ec362-e000-4dae-8b3c-f0d3863c96f5-kube-api-access-fxv5p\") pod \"community-operators-p4zkb\" (UID: \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\") " pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:48 crc kubenswrapper[4931]: I1129 00:48:48.012312 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:48 crc kubenswrapper[4931]: I1129 00:48:48.551616 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4zkb"] Nov 29 00:48:48 crc kubenswrapper[4931]: W1129 00:48:48.571359 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc19ec362_e000_4dae_8b3c_f0d3863c96f5.slice/crio-64fd844be09b3b0446c0822e3db26dbfe8513fe9b56acbba459b1d745700f977 WatchSource:0}: Error finding container 64fd844be09b3b0446c0822e3db26dbfe8513fe9b56acbba459b1d745700f977: Status 404 returned error can't find the container with id 64fd844be09b3b0446c0822e3db26dbfe8513fe9b56acbba459b1d745700f977 Nov 29 00:48:49 crc kubenswrapper[4931]: I1129 00:48:49.568260 4931 generic.go:334] "Generic (PLEG): container finished" podID="c19ec362-e000-4dae-8b3c-f0d3863c96f5" containerID="9e91a92b605ee1455f535b4fc8cc7bf51fd48903ebf19febad000485a653a735" exitCode=0 Nov 29 00:48:49 crc kubenswrapper[4931]: I1129 00:48:49.568406 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4zkb" event={"ID":"c19ec362-e000-4dae-8b3c-f0d3863c96f5","Type":"ContainerDied","Data":"9e91a92b605ee1455f535b4fc8cc7bf51fd48903ebf19febad000485a653a735"} Nov 29 00:48:49 crc kubenswrapper[4931]: I1129 00:48:49.568636 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4zkb" event={"ID":"c19ec362-e000-4dae-8b3c-f0d3863c96f5","Type":"ContainerStarted","Data":"64fd844be09b3b0446c0822e3db26dbfe8513fe9b56acbba459b1d745700f977"} Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.061969 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lrj6v"] Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.067108 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.104186 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrj6v"] Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.251242 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34b613b6-b330-4c23-a97b-ce06bceddf22-utilities\") pod \"redhat-operators-lrj6v\" (UID: \"34b613b6-b330-4c23-a97b-ce06bceddf22\") " pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.251288 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34b613b6-b330-4c23-a97b-ce06bceddf22-catalog-content\") pod \"redhat-operators-lrj6v\" (UID: \"34b613b6-b330-4c23-a97b-ce06bceddf22\") " pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.251329 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgmmm\" (UniqueName: \"kubernetes.io/projected/34b613b6-b330-4c23-a97b-ce06bceddf22-kube-api-access-lgmmm\") pod \"redhat-operators-lrj6v\" (UID: \"34b613b6-b330-4c23-a97b-ce06bceddf22\") " pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.352930 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgmmm\" (UniqueName: \"kubernetes.io/projected/34b613b6-b330-4c23-a97b-ce06bceddf22-kube-api-access-lgmmm\") pod \"redhat-operators-lrj6v\" (UID: \"34b613b6-b330-4c23-a97b-ce06bceddf22\") " pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.353544 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34b613b6-b330-4c23-a97b-ce06bceddf22-utilities\") pod \"redhat-operators-lrj6v\" (UID: \"34b613b6-b330-4c23-a97b-ce06bceddf22\") " pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.353597 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34b613b6-b330-4c23-a97b-ce06bceddf22-catalog-content\") pod \"redhat-operators-lrj6v\" (UID: \"34b613b6-b330-4c23-a97b-ce06bceddf22\") " pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.354399 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34b613b6-b330-4c23-a97b-ce06bceddf22-utilities\") pod \"redhat-operators-lrj6v\" (UID: \"34b613b6-b330-4c23-a97b-ce06bceddf22\") " pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.354421 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34b613b6-b330-4c23-a97b-ce06bceddf22-catalog-content\") pod \"redhat-operators-lrj6v\" (UID: \"34b613b6-b330-4c23-a97b-ce06bceddf22\") " pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.377316 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgmmm\" (UniqueName: \"kubernetes.io/projected/34b613b6-b330-4c23-a97b-ce06bceddf22-kube-api-access-lgmmm\") pod \"redhat-operators-lrj6v\" (UID: \"34b613b6-b330-4c23-a97b-ce06bceddf22\") " pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.427742 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.605370 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4zkb" event={"ID":"c19ec362-e000-4dae-8b3c-f0d3863c96f5","Type":"ContainerStarted","Data":"62016a9989e6fba84ff743b145f09b100c711eb86dab5a0276d23fccad6cd8d2"} Nov 29 00:48:50 crc kubenswrapper[4931]: I1129 00:48:50.908274 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrj6v"] Nov 29 00:48:50 crc kubenswrapper[4931]: W1129 00:48:50.912676 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34b613b6_b330_4c23_a97b_ce06bceddf22.slice/crio-59ba4038914abda5e05f5d44703c8ba8da23be62095cf766c1fa4859a0e3c202 WatchSource:0}: Error finding container 59ba4038914abda5e05f5d44703c8ba8da23be62095cf766c1fa4859a0e3c202: Status 404 returned error can't find the container with id 59ba4038914abda5e05f5d44703c8ba8da23be62095cf766c1fa4859a0e3c202 Nov 29 00:48:51 crc kubenswrapper[4931]: I1129 00:48:51.620440 4931 generic.go:334] "Generic (PLEG): container finished" podID="34b613b6-b330-4c23-a97b-ce06bceddf22" containerID="db1a4cbecf6d29648ac6012da3ba9d4f5421577beb6a56854fbaeaff2cc352fe" exitCode=0 Nov 29 00:48:51 crc kubenswrapper[4931]: I1129 00:48:51.620518 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrj6v" event={"ID":"34b613b6-b330-4c23-a97b-ce06bceddf22","Type":"ContainerDied","Data":"db1a4cbecf6d29648ac6012da3ba9d4f5421577beb6a56854fbaeaff2cc352fe"} Nov 29 00:48:51 crc kubenswrapper[4931]: I1129 00:48:51.620942 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrj6v" event={"ID":"34b613b6-b330-4c23-a97b-ce06bceddf22","Type":"ContainerStarted","Data":"59ba4038914abda5e05f5d44703c8ba8da23be62095cf766c1fa4859a0e3c202"} Nov 29 00:48:51 crc kubenswrapper[4931]: I1129 00:48:51.627270 4931 generic.go:334] "Generic (PLEG): container finished" podID="c19ec362-e000-4dae-8b3c-f0d3863c96f5" containerID="62016a9989e6fba84ff743b145f09b100c711eb86dab5a0276d23fccad6cd8d2" exitCode=0 Nov 29 00:48:51 crc kubenswrapper[4931]: I1129 00:48:51.627324 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4zkb" event={"ID":"c19ec362-e000-4dae-8b3c-f0d3863c96f5","Type":"ContainerDied","Data":"62016a9989e6fba84ff743b145f09b100c711eb86dab5a0276d23fccad6cd8d2"} Nov 29 00:48:52 crc kubenswrapper[4931]: I1129 00:48:52.639835 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4zkb" event={"ID":"c19ec362-e000-4dae-8b3c-f0d3863c96f5","Type":"ContainerStarted","Data":"d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a"} Nov 29 00:48:52 crc kubenswrapper[4931]: I1129 00:48:52.668390 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p4zkb" podStartSLOduration=3.160347011 podStartE2EDuration="5.66836631s" podCreationTimestamp="2025-11-29 00:48:47 +0000 UTC" firstStartedPulling="2025-11-29 00:48:49.570709822 +0000 UTC m=+2942.732603084" lastFinishedPulling="2025-11-29 00:48:52.078729141 +0000 UTC m=+2945.240622383" observedRunningTime="2025-11-29 00:48:52.66731698 +0000 UTC m=+2945.829210262" watchObservedRunningTime="2025-11-29 00:48:52.66836631 +0000 UTC m=+2945.830259572" Nov 29 00:48:53 crc kubenswrapper[4931]: I1129 00:48:53.674645 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrj6v" event={"ID":"34b613b6-b330-4c23-a97b-ce06bceddf22","Type":"ContainerStarted","Data":"785ae564a1545f40f030be00f912a82342e92865eb182a7396d9ea4ada4d2870"} Nov 29 00:48:56 crc kubenswrapper[4931]: I1129 00:48:56.718510 4931 generic.go:334] "Generic (PLEG): container finished" podID="34b613b6-b330-4c23-a97b-ce06bceddf22" containerID="785ae564a1545f40f030be00f912a82342e92865eb182a7396d9ea4ada4d2870" exitCode=0 Nov 29 00:48:56 crc kubenswrapper[4931]: I1129 00:48:56.718620 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrj6v" event={"ID":"34b613b6-b330-4c23-a97b-ce06bceddf22","Type":"ContainerDied","Data":"785ae564a1545f40f030be00f912a82342e92865eb182a7396d9ea4ada4d2870"} Nov 29 00:48:58 crc kubenswrapper[4931]: I1129 00:48:58.012911 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:58 crc kubenswrapper[4931]: I1129 00:48:58.013239 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:58 crc kubenswrapper[4931]: I1129 00:48:58.073107 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:48:58 crc kubenswrapper[4931]: I1129 00:48:58.745845 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrj6v" event={"ID":"34b613b6-b330-4c23-a97b-ce06bceddf22","Type":"ContainerStarted","Data":"9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e"} Nov 29 00:48:58 crc kubenswrapper[4931]: I1129 00:48:58.784744 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lrj6v" podStartSLOduration=2.89623544 podStartE2EDuration="8.784720897s" podCreationTimestamp="2025-11-29 00:48:50 +0000 UTC" firstStartedPulling="2025-11-29 00:48:51.624306892 +0000 UTC m=+2944.786200154" lastFinishedPulling="2025-11-29 00:48:57.512792329 +0000 UTC m=+2950.674685611" observedRunningTime="2025-11-29 00:48:58.772290181 +0000 UTC m=+2951.934183483" watchObservedRunningTime="2025-11-29 00:48:58.784720897 +0000 UTC m=+2951.946614139" Nov 29 00:48:58 crc kubenswrapper[4931]: I1129 00:48:58.838216 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:49:00 crc kubenswrapper[4931]: I1129 00:49:00.257341 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4zkb"] Nov 29 00:49:00 crc kubenswrapper[4931]: I1129 00:49:00.428591 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:49:00 crc kubenswrapper[4931]: I1129 00:49:00.428653 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:49:00 crc kubenswrapper[4931]: I1129 00:49:00.768610 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p4zkb" podUID="c19ec362-e000-4dae-8b3c-f0d3863c96f5" containerName="registry-server" containerID="cri-o://d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a" gracePeriod=2 Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.279866 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.404708 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxv5p\" (UniqueName: \"kubernetes.io/projected/c19ec362-e000-4dae-8b3c-f0d3863c96f5-kube-api-access-fxv5p\") pod \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\" (UID: \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\") " Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.404858 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c19ec362-e000-4dae-8b3c-f0d3863c96f5-catalog-content\") pod \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\" (UID: \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\") " Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.404943 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c19ec362-e000-4dae-8b3c-f0d3863c96f5-utilities\") pod \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\" (UID: \"c19ec362-e000-4dae-8b3c-f0d3863c96f5\") " Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.405729 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c19ec362-e000-4dae-8b3c-f0d3863c96f5-utilities" (OuterVolumeSpecName: "utilities") pod "c19ec362-e000-4dae-8b3c-f0d3863c96f5" (UID: "c19ec362-e000-4dae-8b3c-f0d3863c96f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.418082 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c19ec362-e000-4dae-8b3c-f0d3863c96f5-kube-api-access-fxv5p" (OuterVolumeSpecName: "kube-api-access-fxv5p") pod "c19ec362-e000-4dae-8b3c-f0d3863c96f5" (UID: "c19ec362-e000-4dae-8b3c-f0d3863c96f5"). InnerVolumeSpecName "kube-api-access-fxv5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.463473 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c19ec362-e000-4dae-8b3c-f0d3863c96f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c19ec362-e000-4dae-8b3c-f0d3863c96f5" (UID: "c19ec362-e000-4dae-8b3c-f0d3863c96f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.507881 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxv5p\" (UniqueName: \"kubernetes.io/projected/c19ec362-e000-4dae-8b3c-f0d3863c96f5-kube-api-access-fxv5p\") on node \"crc\" DevicePath \"\"" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.507931 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c19ec362-e000-4dae-8b3c-f0d3863c96f5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.507952 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c19ec362-e000-4dae-8b3c-f0d3863c96f5-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.512426 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lrj6v" podUID="34b613b6-b330-4c23-a97b-ce06bceddf22" containerName="registry-server" probeResult="failure" output=< Nov 29 00:49:01 crc kubenswrapper[4931]: timeout: failed to connect service ":50051" within 1s Nov 29 00:49:01 crc kubenswrapper[4931]: > Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.784040 4931 generic.go:334] "Generic (PLEG): container finished" podID="c19ec362-e000-4dae-8b3c-f0d3863c96f5" containerID="d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a" exitCode=0 Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.784106 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4zkb" event={"ID":"c19ec362-e000-4dae-8b3c-f0d3863c96f5","Type":"ContainerDied","Data":"d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a"} Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.784137 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4zkb" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.784167 4931 scope.go:117] "RemoveContainer" containerID="d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.784149 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4zkb" event={"ID":"c19ec362-e000-4dae-8b3c-f0d3863c96f5","Type":"ContainerDied","Data":"64fd844be09b3b0446c0822e3db26dbfe8513fe9b56acbba459b1d745700f977"} Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.818437 4931 scope.go:117] "RemoveContainer" containerID="62016a9989e6fba84ff743b145f09b100c711eb86dab5a0276d23fccad6cd8d2" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.841701 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4zkb"] Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.852167 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p4zkb"] Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.929240 4931 scope.go:117] "RemoveContainer" containerID="9e91a92b605ee1455f535b4fc8cc7bf51fd48903ebf19febad000485a653a735" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.972257 4931 scope.go:117] "RemoveContainer" containerID="d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a" Nov 29 00:49:01 crc kubenswrapper[4931]: E1129 00:49:01.972984 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a\": container with ID starting with d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a not found: ID does not exist" containerID="d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.973010 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a"} err="failed to get container status \"d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a\": rpc error: code = NotFound desc = could not find container \"d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a\": container with ID starting with d9048749a876bf188384520be72985db9df971185426061441e12f843ef6392a not found: ID does not exist" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.973036 4931 scope.go:117] "RemoveContainer" containerID="62016a9989e6fba84ff743b145f09b100c711eb86dab5a0276d23fccad6cd8d2" Nov 29 00:49:01 crc kubenswrapper[4931]: E1129 00:49:01.973424 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62016a9989e6fba84ff743b145f09b100c711eb86dab5a0276d23fccad6cd8d2\": container with ID starting with 62016a9989e6fba84ff743b145f09b100c711eb86dab5a0276d23fccad6cd8d2 not found: ID does not exist" containerID="62016a9989e6fba84ff743b145f09b100c711eb86dab5a0276d23fccad6cd8d2" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.973478 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62016a9989e6fba84ff743b145f09b100c711eb86dab5a0276d23fccad6cd8d2"} err="failed to get container status \"62016a9989e6fba84ff743b145f09b100c711eb86dab5a0276d23fccad6cd8d2\": rpc error: code = NotFound desc = could not find container \"62016a9989e6fba84ff743b145f09b100c711eb86dab5a0276d23fccad6cd8d2\": container with ID starting with 62016a9989e6fba84ff743b145f09b100c711eb86dab5a0276d23fccad6cd8d2 not found: ID does not exist" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.973515 4931 scope.go:117] "RemoveContainer" containerID="9e91a92b605ee1455f535b4fc8cc7bf51fd48903ebf19febad000485a653a735" Nov 29 00:49:01 crc kubenswrapper[4931]: E1129 00:49:01.973780 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e91a92b605ee1455f535b4fc8cc7bf51fd48903ebf19febad000485a653a735\": container with ID starting with 9e91a92b605ee1455f535b4fc8cc7bf51fd48903ebf19febad000485a653a735 not found: ID does not exist" containerID="9e91a92b605ee1455f535b4fc8cc7bf51fd48903ebf19febad000485a653a735" Nov 29 00:49:01 crc kubenswrapper[4931]: I1129 00:49:01.973822 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e91a92b605ee1455f535b4fc8cc7bf51fd48903ebf19febad000485a653a735"} err="failed to get container status \"9e91a92b605ee1455f535b4fc8cc7bf51fd48903ebf19febad000485a653a735\": rpc error: code = NotFound desc = could not find container \"9e91a92b605ee1455f535b4fc8cc7bf51fd48903ebf19febad000485a653a735\": container with ID starting with 9e91a92b605ee1455f535b4fc8cc7bf51fd48903ebf19febad000485a653a735 not found: ID does not exist" Nov 29 00:49:03 crc kubenswrapper[4931]: I1129 00:49:03.227651 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c19ec362-e000-4dae-8b3c-f0d3863c96f5" path="/var/lib/kubelet/pods/c19ec362-e000-4dae-8b3c-f0d3863c96f5/volumes" Nov 29 00:49:10 crc kubenswrapper[4931]: I1129 00:49:10.483012 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:49:10 crc kubenswrapper[4931]: I1129 00:49:10.549782 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:49:10 crc kubenswrapper[4931]: I1129 00:49:10.747084 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrj6v"] Nov 29 00:49:11 crc kubenswrapper[4931]: I1129 00:49:11.890933 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lrj6v" podUID="34b613b6-b330-4c23-a97b-ce06bceddf22" containerName="registry-server" containerID="cri-o://9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e" gracePeriod=2 Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.400334 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.474192 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34b613b6-b330-4c23-a97b-ce06bceddf22-catalog-content\") pod \"34b613b6-b330-4c23-a97b-ce06bceddf22\" (UID: \"34b613b6-b330-4c23-a97b-ce06bceddf22\") " Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.474244 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34b613b6-b330-4c23-a97b-ce06bceddf22-utilities\") pod \"34b613b6-b330-4c23-a97b-ce06bceddf22\" (UID: \"34b613b6-b330-4c23-a97b-ce06bceddf22\") " Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.474306 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgmmm\" (UniqueName: \"kubernetes.io/projected/34b613b6-b330-4c23-a97b-ce06bceddf22-kube-api-access-lgmmm\") pod \"34b613b6-b330-4c23-a97b-ce06bceddf22\" (UID: \"34b613b6-b330-4c23-a97b-ce06bceddf22\") " Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.481012 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34b613b6-b330-4c23-a97b-ce06bceddf22-utilities" (OuterVolumeSpecName: "utilities") pod "34b613b6-b330-4c23-a97b-ce06bceddf22" (UID: "34b613b6-b330-4c23-a97b-ce06bceddf22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.482976 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34b613b6-b330-4c23-a97b-ce06bceddf22-kube-api-access-lgmmm" (OuterVolumeSpecName: "kube-api-access-lgmmm") pod "34b613b6-b330-4c23-a97b-ce06bceddf22" (UID: "34b613b6-b330-4c23-a97b-ce06bceddf22"). InnerVolumeSpecName "kube-api-access-lgmmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.578496 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34b613b6-b330-4c23-a97b-ce06bceddf22-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.578668 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgmmm\" (UniqueName: \"kubernetes.io/projected/34b613b6-b330-4c23-a97b-ce06bceddf22-kube-api-access-lgmmm\") on node \"crc\" DevicePath \"\"" Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.585740 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34b613b6-b330-4c23-a97b-ce06bceddf22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34b613b6-b330-4c23-a97b-ce06bceddf22" (UID: "34b613b6-b330-4c23-a97b-ce06bceddf22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.680154 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34b613b6-b330-4c23-a97b-ce06bceddf22-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.905460 4931 generic.go:334] "Generic (PLEG): container finished" podID="34b613b6-b330-4c23-a97b-ce06bceddf22" containerID="9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e" exitCode=0 Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.905741 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrj6v" event={"ID":"34b613b6-b330-4c23-a97b-ce06bceddf22","Type":"ContainerDied","Data":"9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e"} Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.905873 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrj6v" Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.906002 4931 scope.go:117] "RemoveContainer" containerID="9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e" Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.905974 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrj6v" event={"ID":"34b613b6-b330-4c23-a97b-ce06bceddf22","Type":"ContainerDied","Data":"59ba4038914abda5e05f5d44703c8ba8da23be62095cf766c1fa4859a0e3c202"} Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.948430 4931 scope.go:117] "RemoveContainer" containerID="785ae564a1545f40f030be00f912a82342e92865eb182a7396d9ea4ada4d2870" Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.972747 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrj6v"] Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.985443 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lrj6v"] Nov 29 00:49:12 crc kubenswrapper[4931]: I1129 00:49:12.988516 4931 scope.go:117] "RemoveContainer" containerID="db1a4cbecf6d29648ac6012da3ba9d4f5421577beb6a56854fbaeaff2cc352fe" Nov 29 00:49:13 crc kubenswrapper[4931]: I1129 00:49:13.034643 4931 scope.go:117] "RemoveContainer" containerID="9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e" Nov 29 00:49:13 crc kubenswrapper[4931]: E1129 00:49:13.035157 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e\": container with ID starting with 9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e not found: ID does not exist" containerID="9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e" Nov 29 00:49:13 crc kubenswrapper[4931]: I1129 00:49:13.035204 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e"} err="failed to get container status \"9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e\": rpc error: code = NotFound desc = could not find container \"9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e\": container with ID starting with 9ece96c369a4508ef4f5f0b987783324e2330c709ecf11eb0316a75201575d4e not found: ID does not exist" Nov 29 00:49:13 crc kubenswrapper[4931]: I1129 00:49:13.035236 4931 scope.go:117] "RemoveContainer" containerID="785ae564a1545f40f030be00f912a82342e92865eb182a7396d9ea4ada4d2870" Nov 29 00:49:13 crc kubenswrapper[4931]: E1129 00:49:13.035863 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"785ae564a1545f40f030be00f912a82342e92865eb182a7396d9ea4ada4d2870\": container with ID starting with 785ae564a1545f40f030be00f912a82342e92865eb182a7396d9ea4ada4d2870 not found: ID does not exist" containerID="785ae564a1545f40f030be00f912a82342e92865eb182a7396d9ea4ada4d2870" Nov 29 00:49:13 crc kubenswrapper[4931]: I1129 00:49:13.035910 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"785ae564a1545f40f030be00f912a82342e92865eb182a7396d9ea4ada4d2870"} err="failed to get container status \"785ae564a1545f40f030be00f912a82342e92865eb182a7396d9ea4ada4d2870\": rpc error: code = NotFound desc = could not find container \"785ae564a1545f40f030be00f912a82342e92865eb182a7396d9ea4ada4d2870\": container with ID starting with 785ae564a1545f40f030be00f912a82342e92865eb182a7396d9ea4ada4d2870 not found: ID does not exist" Nov 29 00:49:13 crc kubenswrapper[4931]: I1129 00:49:13.035937 4931 scope.go:117] "RemoveContainer" containerID="db1a4cbecf6d29648ac6012da3ba9d4f5421577beb6a56854fbaeaff2cc352fe" Nov 29 00:49:13 crc kubenswrapper[4931]: E1129 00:49:13.036284 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db1a4cbecf6d29648ac6012da3ba9d4f5421577beb6a56854fbaeaff2cc352fe\": container with ID starting with db1a4cbecf6d29648ac6012da3ba9d4f5421577beb6a56854fbaeaff2cc352fe not found: ID does not exist" containerID="db1a4cbecf6d29648ac6012da3ba9d4f5421577beb6a56854fbaeaff2cc352fe" Nov 29 00:49:13 crc kubenswrapper[4931]: I1129 00:49:13.036317 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db1a4cbecf6d29648ac6012da3ba9d4f5421577beb6a56854fbaeaff2cc352fe"} err="failed to get container status \"db1a4cbecf6d29648ac6012da3ba9d4f5421577beb6a56854fbaeaff2cc352fe\": rpc error: code = NotFound desc = could not find container \"db1a4cbecf6d29648ac6012da3ba9d4f5421577beb6a56854fbaeaff2cc352fe\": container with ID starting with db1a4cbecf6d29648ac6012da3ba9d4f5421577beb6a56854fbaeaff2cc352fe not found: ID does not exist" Nov 29 00:49:13 crc kubenswrapper[4931]: I1129 00:49:13.228137 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34b613b6-b330-4c23-a97b-ce06bceddf22" path="/var/lib/kubelet/pods/34b613b6-b330-4c23-a97b-ce06bceddf22/volumes" Nov 29 00:50:53 crc kubenswrapper[4931]: I1129 00:50:53.734855 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:50:53 crc kubenswrapper[4931]: I1129 00:50:53.735465 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:51:07 crc kubenswrapper[4931]: I1129 00:51:07.274982 4931 generic.go:334] "Generic (PLEG): container finished" podID="6a410480-c08c-4a1a-b3a4-6d2372d6bb2a" containerID="07cea9ced249a9827f6811ad7c6901bd203497cc31b9dc8f5c00896a5789ee2b" exitCode=0 Nov 29 00:51:07 crc kubenswrapper[4931]: I1129 00:51:07.275119 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" event={"ID":"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a","Type":"ContainerDied","Data":"07cea9ced249a9827f6811ad7c6901bd203497cc31b9dc8f5c00896a5789ee2b"} Nov 29 00:51:08 crc kubenswrapper[4931]: I1129 00:51:08.874640 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:51:08 crc kubenswrapper[4931]: I1129 00:51:08.981456 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ssh-key\") pod \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " Nov 29 00:51:08 crc kubenswrapper[4931]: I1129 00:51:08.982295 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkt66\" (UniqueName: \"kubernetes.io/projected/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-kube-api-access-tkt66\") pod \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " Nov 29 00:51:08 crc kubenswrapper[4931]: I1129 00:51:08.982349 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-telemetry-combined-ca-bundle\") pod \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " Nov 29 00:51:08 crc kubenswrapper[4931]: I1129 00:51:08.982408 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-2\") pod \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " Nov 29 00:51:08 crc kubenswrapper[4931]: I1129 00:51:08.982521 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-0\") pod \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " Nov 29 00:51:08 crc kubenswrapper[4931]: I1129 00:51:08.982692 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-1\") pod \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " Nov 29 00:51:08 crc kubenswrapper[4931]: I1129 00:51:08.982778 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-inventory\") pod \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\" (UID: \"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a\") " Nov 29 00:51:08 crc kubenswrapper[4931]: I1129 00:51:08.992219 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a" (UID: "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:51:08 crc kubenswrapper[4931]: I1129 00:51:08.993290 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-kube-api-access-tkt66" (OuterVolumeSpecName: "kube-api-access-tkt66") pod "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a" (UID: "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a"). InnerVolumeSpecName "kube-api-access-tkt66". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.034221 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a" (UID: "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.036379 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a" (UID: "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.037432 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-inventory" (OuterVolumeSpecName: "inventory") pod "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a" (UID: "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.046213 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a" (UID: "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.046235 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a" (UID: "6a410480-c08c-4a1a-b3a4-6d2372d6bb2a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.084936 4931 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.084974 4931 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-inventory\") on node \"crc\" DevicePath \"\"" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.084986 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.084998 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkt66\" (UniqueName: \"kubernetes.io/projected/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-kube-api-access-tkt66\") on node \"crc\" DevicePath \"\"" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.085011 4931 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.085022 4931 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.085033 4931 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/6a410480-c08c-4a1a-b3a4-6d2372d6bb2a-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.310304 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" event={"ID":"6a410480-c08c-4a1a-b3a4-6d2372d6bb2a","Type":"ContainerDied","Data":"7be66c7a2656298d93f7fdd522b9202e98c7d90d7983dab23b7502908b715bcb"} Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.310348 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7be66c7a2656298d93f7fdd522b9202e98c7d90d7983dab23b7502908b715bcb" Nov 29 00:51:09 crc kubenswrapper[4931]: I1129 00:51:09.310367 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj" Nov 29 00:51:23 crc kubenswrapper[4931]: I1129 00:51:23.734803 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:51:23 crc kubenswrapper[4931]: I1129 00:51:23.735501 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:51:53 crc kubenswrapper[4931]: I1129 00:51:53.734698 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:51:53 crc kubenswrapper[4931]: I1129 00:51:53.735148 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:51:53 crc kubenswrapper[4931]: I1129 00:51:53.735196 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 00:51:53 crc kubenswrapper[4931]: I1129 00:51:53.735908 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 00:51:53 crc kubenswrapper[4931]: I1129 00:51:53.735961 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" gracePeriod=600 Nov 29 00:51:53 crc kubenswrapper[4931]: E1129 00:51:53.859831 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:51:53 crc kubenswrapper[4931]: I1129 00:51:53.869982 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" exitCode=0 Nov 29 00:51:53 crc kubenswrapper[4931]: I1129 00:51:53.870033 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251"} Nov 29 00:51:53 crc kubenswrapper[4931]: I1129 00:51:53.870082 4931 scope.go:117] "RemoveContainer" containerID="3d459f7554777705e26d66be3e626f25d6b92931660a2f3ad9dff8fdaba9654f" Nov 29 00:51:53 crc kubenswrapper[4931]: I1129 00:51:53.870741 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:51:53 crc kubenswrapper[4931]: E1129 00:51:53.871008 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.330407 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 29 00:52:02 crc kubenswrapper[4931]: E1129 00:52:02.331907 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a410480-c08c-4a1a-b3a4-6d2372d6bb2a" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.331938 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a410480-c08c-4a1a-b3a4-6d2372d6bb2a" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 29 00:52:02 crc kubenswrapper[4931]: E1129 00:52:02.331964 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19ec362-e000-4dae-8b3c-f0d3863c96f5" containerName="registry-server" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.331979 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19ec362-e000-4dae-8b3c-f0d3863c96f5" containerName="registry-server" Nov 29 00:52:02 crc kubenswrapper[4931]: E1129 00:52:02.331999 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19ec362-e000-4dae-8b3c-f0d3863c96f5" containerName="extract-content" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.332011 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19ec362-e000-4dae-8b3c-f0d3863c96f5" containerName="extract-content" Nov 29 00:52:02 crc kubenswrapper[4931]: E1129 00:52:02.332027 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19ec362-e000-4dae-8b3c-f0d3863c96f5" containerName="extract-utilities" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.332039 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19ec362-e000-4dae-8b3c-f0d3863c96f5" containerName="extract-utilities" Nov 29 00:52:02 crc kubenswrapper[4931]: E1129 00:52:02.332067 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b613b6-b330-4c23-a97b-ce06bceddf22" containerName="extract-content" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.332078 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b613b6-b330-4c23-a97b-ce06bceddf22" containerName="extract-content" Nov 29 00:52:02 crc kubenswrapper[4931]: E1129 00:52:02.332101 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b613b6-b330-4c23-a97b-ce06bceddf22" containerName="registry-server" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.332112 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b613b6-b330-4c23-a97b-ce06bceddf22" containerName="registry-server" Nov 29 00:52:02 crc kubenswrapper[4931]: E1129 00:52:02.332142 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b613b6-b330-4c23-a97b-ce06bceddf22" containerName="extract-utilities" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.332153 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b613b6-b330-4c23-a97b-ce06bceddf22" containerName="extract-utilities" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.332494 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="c19ec362-e000-4dae-8b3c-f0d3863c96f5" containerName="registry-server" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.332529 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="34b613b6-b330-4c23-a97b-ce06bceddf22" containerName="registry-server" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.332579 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a410480-c08c-4a1a-b3a4-6d2372d6bb2a" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.333691 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.336030 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.336578 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.336817 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gnq5t" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.336960 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.356104 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.451165 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/16c927ea-8729-4da8-8893-c23a53d85c5d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.451242 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16c927ea-8729-4da8-8893-c23a53d85c5d-config-data\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.451281 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.451794 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.451958 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/16c927ea-8729-4da8-8893-c23a53d85c5d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.452037 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msmzv\" (UniqueName: \"kubernetes.io/projected/16c927ea-8729-4da8-8893-c23a53d85c5d-kube-api-access-msmzv\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.452088 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.452695 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/16c927ea-8729-4da8-8893-c23a53d85c5d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.452801 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.554367 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/16c927ea-8729-4da8-8893-c23a53d85c5d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.554428 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.554478 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/16c927ea-8729-4da8-8893-c23a53d85c5d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.554520 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16c927ea-8729-4da8-8893-c23a53d85c5d-config-data\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.554550 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.554626 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.554655 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/16c927ea-8729-4da8-8893-c23a53d85c5d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.554681 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msmzv\" (UniqueName: \"kubernetes.io/projected/16c927ea-8729-4da8-8893-c23a53d85c5d-kube-api-access-msmzv\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.554704 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.555776 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/16c927ea-8729-4da8-8893-c23a53d85c5d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.555789 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/16c927ea-8729-4da8-8893-c23a53d85c5d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.557462 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16c927ea-8729-4da8-8893-c23a53d85c5d-config-data\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.557709 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.560094 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/16c927ea-8729-4da8-8893-c23a53d85c5d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.567756 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.568292 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.571652 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.619555 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msmzv\" (UniqueName: \"kubernetes.io/projected/16c927ea-8729-4da8-8893-c23a53d85c5d-kube-api-access-msmzv\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.695132 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " pod="openstack/tempest-tests-tempest" Nov 29 00:52:02 crc kubenswrapper[4931]: I1129 00:52:02.966958 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 29 00:52:03 crc kubenswrapper[4931]: I1129 00:52:03.328091 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 29 00:52:03 crc kubenswrapper[4931]: W1129 00:52:03.336533 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16c927ea_8729_4da8_8893_c23a53d85c5d.slice/crio-4e76f9057cfa66cfee9ceb419a2154295796ac12b981805e6dea9893373c2abb WatchSource:0}: Error finding container 4e76f9057cfa66cfee9ceb419a2154295796ac12b981805e6dea9893373c2abb: Status 404 returned error can't find the container with id 4e76f9057cfa66cfee9ceb419a2154295796ac12b981805e6dea9893373c2abb Nov 29 00:52:03 crc kubenswrapper[4931]: I1129 00:52:03.339239 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 00:52:04 crc kubenswrapper[4931]: I1129 00:52:04.004520 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"16c927ea-8729-4da8-8893-c23a53d85c5d","Type":"ContainerStarted","Data":"4e76f9057cfa66cfee9ceb419a2154295796ac12b981805e6dea9893373c2abb"} Nov 29 00:52:05 crc kubenswrapper[4931]: I1129 00:52:05.212895 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:52:05 crc kubenswrapper[4931]: E1129 00:52:05.213532 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:52:18 crc kubenswrapper[4931]: I1129 00:52:18.212586 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:52:18 crc kubenswrapper[4931]: E1129 00:52:18.213544 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:52:33 crc kubenswrapper[4931]: I1129 00:52:33.213211 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:52:33 crc kubenswrapper[4931]: E1129 00:52:33.214574 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:52:37 crc kubenswrapper[4931]: E1129 00:52:37.971621 4931 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 29 00:52:37 crc kubenswrapper[4931]: E1129 00:52:37.972129 4931 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-msmzv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(16c927ea-8729-4da8-8893-c23a53d85c5d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 29 00:52:37 crc kubenswrapper[4931]: E1129 00:52:37.973363 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="16c927ea-8729-4da8-8893-c23a53d85c5d" Nov 29 00:52:38 crc kubenswrapper[4931]: E1129 00:52:38.427073 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="16c927ea-8729-4da8-8893-c23a53d85c5d" Nov 29 00:52:44 crc kubenswrapper[4931]: I1129 00:52:44.213217 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:52:44 crc kubenswrapper[4931]: E1129 00:52:44.213733 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:52:50 crc kubenswrapper[4931]: I1129 00:52:50.713785 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 29 00:52:52 crc kubenswrapper[4931]: I1129 00:52:52.615866 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"16c927ea-8729-4da8-8893-c23a53d85c5d","Type":"ContainerStarted","Data":"29f062c2ceeb95bc261f68a0d641884366dd6e33146c7a8b6627b1e7a7797c40"} Nov 29 00:52:52 crc kubenswrapper[4931]: I1129 00:52:52.640264 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.268949819 podStartE2EDuration="51.640249275s" podCreationTimestamp="2025-11-29 00:52:01 +0000 UTC" firstStartedPulling="2025-11-29 00:52:03.338949852 +0000 UTC m=+3136.500843104" lastFinishedPulling="2025-11-29 00:52:50.710249288 +0000 UTC m=+3183.872142560" observedRunningTime="2025-11-29 00:52:52.639408751 +0000 UTC m=+3185.801302033" watchObservedRunningTime="2025-11-29 00:52:52.640249275 +0000 UTC m=+3185.802142507" Nov 29 00:52:56 crc kubenswrapper[4931]: I1129 00:52:56.212876 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:52:56 crc kubenswrapper[4931]: E1129 00:52:56.213951 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:53:09 crc kubenswrapper[4931]: I1129 00:53:09.212219 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:53:09 crc kubenswrapper[4931]: E1129 00:53:09.213461 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:53:20 crc kubenswrapper[4931]: I1129 00:53:20.213099 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:53:20 crc kubenswrapper[4931]: E1129 00:53:20.214155 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:53:31 crc kubenswrapper[4931]: I1129 00:53:31.214063 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:53:31 crc kubenswrapper[4931]: E1129 00:53:31.215657 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:53:44 crc kubenswrapper[4931]: I1129 00:53:44.213534 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:53:44 crc kubenswrapper[4931]: E1129 00:53:44.214563 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:53:59 crc kubenswrapper[4931]: I1129 00:53:59.213178 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:53:59 crc kubenswrapper[4931]: E1129 00:53:59.214378 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:54:12 crc kubenswrapper[4931]: I1129 00:54:12.238085 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:54:12 crc kubenswrapper[4931]: E1129 00:54:12.238865 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:54:27 crc kubenswrapper[4931]: I1129 00:54:27.226065 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:54:27 crc kubenswrapper[4931]: E1129 00:54:27.227323 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:54:42 crc kubenswrapper[4931]: I1129 00:54:42.212532 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:54:42 crc kubenswrapper[4931]: E1129 00:54:42.213309 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:54:57 crc kubenswrapper[4931]: I1129 00:54:57.212519 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:54:57 crc kubenswrapper[4931]: E1129 00:54:57.213202 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:55:10 crc kubenswrapper[4931]: I1129 00:55:10.213279 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:55:10 crc kubenswrapper[4931]: E1129 00:55:10.214376 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:55:16 crc kubenswrapper[4931]: I1129 00:55:16.957940 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-56c8n"] Nov 29 00:55:16 crc kubenswrapper[4931]: I1129 00:55:16.960943 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:16 crc kubenswrapper[4931]: I1129 00:55:16.979877 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-56c8n"] Nov 29 00:55:17 crc kubenswrapper[4931]: I1129 00:55:17.055716 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19cbd316-da53-4f3c-a17a-418c297a74b7-catalog-content\") pod \"redhat-marketplace-56c8n\" (UID: \"19cbd316-da53-4f3c-a17a-418c297a74b7\") " pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:17 crc kubenswrapper[4931]: I1129 00:55:17.055792 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19cbd316-da53-4f3c-a17a-418c297a74b7-utilities\") pod \"redhat-marketplace-56c8n\" (UID: \"19cbd316-da53-4f3c-a17a-418c297a74b7\") " pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:17 crc kubenswrapper[4931]: I1129 00:55:17.055921 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2xzx\" (UniqueName: \"kubernetes.io/projected/19cbd316-da53-4f3c-a17a-418c297a74b7-kube-api-access-v2xzx\") pod \"redhat-marketplace-56c8n\" (UID: \"19cbd316-da53-4f3c-a17a-418c297a74b7\") " pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:17 crc kubenswrapper[4931]: I1129 00:55:17.158579 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19cbd316-da53-4f3c-a17a-418c297a74b7-catalog-content\") pod \"redhat-marketplace-56c8n\" (UID: \"19cbd316-da53-4f3c-a17a-418c297a74b7\") " pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:17 crc kubenswrapper[4931]: I1129 00:55:17.158669 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19cbd316-da53-4f3c-a17a-418c297a74b7-utilities\") pod \"redhat-marketplace-56c8n\" (UID: \"19cbd316-da53-4f3c-a17a-418c297a74b7\") " pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:17 crc kubenswrapper[4931]: I1129 00:55:17.158705 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2xzx\" (UniqueName: \"kubernetes.io/projected/19cbd316-da53-4f3c-a17a-418c297a74b7-kube-api-access-v2xzx\") pod \"redhat-marketplace-56c8n\" (UID: \"19cbd316-da53-4f3c-a17a-418c297a74b7\") " pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:17 crc kubenswrapper[4931]: I1129 00:55:17.159800 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19cbd316-da53-4f3c-a17a-418c297a74b7-catalog-content\") pod \"redhat-marketplace-56c8n\" (UID: \"19cbd316-da53-4f3c-a17a-418c297a74b7\") " pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:17 crc kubenswrapper[4931]: I1129 00:55:17.159895 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19cbd316-da53-4f3c-a17a-418c297a74b7-utilities\") pod \"redhat-marketplace-56c8n\" (UID: \"19cbd316-da53-4f3c-a17a-418c297a74b7\") " pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:17 crc kubenswrapper[4931]: I1129 00:55:17.184863 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2xzx\" (UniqueName: \"kubernetes.io/projected/19cbd316-da53-4f3c-a17a-418c297a74b7-kube-api-access-v2xzx\") pod \"redhat-marketplace-56c8n\" (UID: \"19cbd316-da53-4f3c-a17a-418c297a74b7\") " pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:17 crc kubenswrapper[4931]: I1129 00:55:17.287090 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:17 crc kubenswrapper[4931]: I1129 00:55:17.766159 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-56c8n"] Nov 29 00:55:18 crc kubenswrapper[4931]: I1129 00:55:18.446088 4931 generic.go:334] "Generic (PLEG): container finished" podID="19cbd316-da53-4f3c-a17a-418c297a74b7" containerID="4c039ea56671441919d198cacf9a3a5f441bb04ff71c6660685d90e80193b2a1" exitCode=0 Nov 29 00:55:18 crc kubenswrapper[4931]: I1129 00:55:18.446166 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56c8n" event={"ID":"19cbd316-da53-4f3c-a17a-418c297a74b7","Type":"ContainerDied","Data":"4c039ea56671441919d198cacf9a3a5f441bb04ff71c6660685d90e80193b2a1"} Nov 29 00:55:18 crc kubenswrapper[4931]: I1129 00:55:18.446596 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56c8n" event={"ID":"19cbd316-da53-4f3c-a17a-418c297a74b7","Type":"ContainerStarted","Data":"95651b50c382e883af12fa1463b666ea040780266103e2303c50ed203045fb8e"} Nov 29 00:55:19 crc kubenswrapper[4931]: I1129 00:55:19.461041 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56c8n" event={"ID":"19cbd316-da53-4f3c-a17a-418c297a74b7","Type":"ContainerStarted","Data":"54822ce5f3b2bf8c5856f782cd0169e972fa7ecd2314451c98e0cb9a37209913"} Nov 29 00:55:20 crc kubenswrapper[4931]: I1129 00:55:20.477396 4931 generic.go:334] "Generic (PLEG): container finished" podID="19cbd316-da53-4f3c-a17a-418c297a74b7" containerID="54822ce5f3b2bf8c5856f782cd0169e972fa7ecd2314451c98e0cb9a37209913" exitCode=0 Nov 29 00:55:20 crc kubenswrapper[4931]: I1129 00:55:20.477514 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56c8n" event={"ID":"19cbd316-da53-4f3c-a17a-418c297a74b7","Type":"ContainerDied","Data":"54822ce5f3b2bf8c5856f782cd0169e972fa7ecd2314451c98e0cb9a37209913"} Nov 29 00:55:21 crc kubenswrapper[4931]: I1129 00:55:21.490438 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56c8n" event={"ID":"19cbd316-da53-4f3c-a17a-418c297a74b7","Type":"ContainerStarted","Data":"1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5"} Nov 29 00:55:24 crc kubenswrapper[4931]: I1129 00:55:24.212396 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:55:24 crc kubenswrapper[4931]: E1129 00:55:24.213203 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:55:27 crc kubenswrapper[4931]: I1129 00:55:27.288431 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:27 crc kubenswrapper[4931]: I1129 00:55:27.289291 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:27 crc kubenswrapper[4931]: I1129 00:55:27.373260 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:27 crc kubenswrapper[4931]: I1129 00:55:27.409199 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-56c8n" podStartSLOduration=8.820193447 podStartE2EDuration="11.409179133s" podCreationTimestamp="2025-11-29 00:55:16 +0000 UTC" firstStartedPulling="2025-11-29 00:55:18.448695137 +0000 UTC m=+3331.610588369" lastFinishedPulling="2025-11-29 00:55:21.037680823 +0000 UTC m=+3334.199574055" observedRunningTime="2025-11-29 00:55:21.515859817 +0000 UTC m=+3334.677753079" watchObservedRunningTime="2025-11-29 00:55:27.409179133 +0000 UTC m=+3340.571072365" Nov 29 00:55:27 crc kubenswrapper[4931]: I1129 00:55:27.632400 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:27 crc kubenswrapper[4931]: I1129 00:55:27.712691 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-56c8n"] Nov 29 00:55:29 crc kubenswrapper[4931]: I1129 00:55:29.588888 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-56c8n" podUID="19cbd316-da53-4f3c-a17a-418c297a74b7" containerName="registry-server" containerID="cri-o://1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5" gracePeriod=2 Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.147711 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.269936 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19cbd316-da53-4f3c-a17a-418c297a74b7-utilities\") pod \"19cbd316-da53-4f3c-a17a-418c297a74b7\" (UID: \"19cbd316-da53-4f3c-a17a-418c297a74b7\") " Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.270067 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2xzx\" (UniqueName: \"kubernetes.io/projected/19cbd316-da53-4f3c-a17a-418c297a74b7-kube-api-access-v2xzx\") pod \"19cbd316-da53-4f3c-a17a-418c297a74b7\" (UID: \"19cbd316-da53-4f3c-a17a-418c297a74b7\") " Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.271475 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19cbd316-da53-4f3c-a17a-418c297a74b7-catalog-content\") pod \"19cbd316-da53-4f3c-a17a-418c297a74b7\" (UID: \"19cbd316-da53-4f3c-a17a-418c297a74b7\") " Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.280303 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19cbd316-da53-4f3c-a17a-418c297a74b7-kube-api-access-v2xzx" (OuterVolumeSpecName: "kube-api-access-v2xzx") pod "19cbd316-da53-4f3c-a17a-418c297a74b7" (UID: "19cbd316-da53-4f3c-a17a-418c297a74b7"). InnerVolumeSpecName "kube-api-access-v2xzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.289229 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19cbd316-da53-4f3c-a17a-418c297a74b7-utilities" (OuterVolumeSpecName: "utilities") pod "19cbd316-da53-4f3c-a17a-418c297a74b7" (UID: "19cbd316-da53-4f3c-a17a-418c297a74b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.320339 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19cbd316-da53-4f3c-a17a-418c297a74b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19cbd316-da53-4f3c-a17a-418c297a74b7" (UID: "19cbd316-da53-4f3c-a17a-418c297a74b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.375884 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2xzx\" (UniqueName: \"kubernetes.io/projected/19cbd316-da53-4f3c-a17a-418c297a74b7-kube-api-access-v2xzx\") on node \"crc\" DevicePath \"\"" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.375915 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19cbd316-da53-4f3c-a17a-418c297a74b7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.375929 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19cbd316-da53-4f3c-a17a-418c297a74b7-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.604152 4931 generic.go:334] "Generic (PLEG): container finished" podID="19cbd316-da53-4f3c-a17a-418c297a74b7" containerID="1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5" exitCode=0 Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.604208 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56c8n" event={"ID":"19cbd316-da53-4f3c-a17a-418c297a74b7","Type":"ContainerDied","Data":"1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5"} Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.604282 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56c8n" event={"ID":"19cbd316-da53-4f3c-a17a-418c297a74b7","Type":"ContainerDied","Data":"95651b50c382e883af12fa1463b666ea040780266103e2303c50ed203045fb8e"} Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.604305 4931 scope.go:117] "RemoveContainer" containerID="1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.604244 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56c8n" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.649753 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-56c8n"] Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.662777 4931 scope.go:117] "RemoveContainer" containerID="54822ce5f3b2bf8c5856f782cd0169e972fa7ecd2314451c98e0cb9a37209913" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.664260 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-56c8n"] Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.689614 4931 scope.go:117] "RemoveContainer" containerID="4c039ea56671441919d198cacf9a3a5f441bb04ff71c6660685d90e80193b2a1" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.751825 4931 scope.go:117] "RemoveContainer" containerID="1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5" Nov 29 00:55:30 crc kubenswrapper[4931]: E1129 00:55:30.752781 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5\": container with ID starting with 1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5 not found: ID does not exist" containerID="1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.752931 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5"} err="failed to get container status \"1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5\": rpc error: code = NotFound desc = could not find container \"1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5\": container with ID starting with 1abf189f1b9c1580b82c38d689d6a0e46d3c4c8f74f3827b947adf3a245eafd5 not found: ID does not exist" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.752977 4931 scope.go:117] "RemoveContainer" containerID="54822ce5f3b2bf8c5856f782cd0169e972fa7ecd2314451c98e0cb9a37209913" Nov 29 00:55:30 crc kubenswrapper[4931]: E1129 00:55:30.753341 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54822ce5f3b2bf8c5856f782cd0169e972fa7ecd2314451c98e0cb9a37209913\": container with ID starting with 54822ce5f3b2bf8c5856f782cd0169e972fa7ecd2314451c98e0cb9a37209913 not found: ID does not exist" containerID="54822ce5f3b2bf8c5856f782cd0169e972fa7ecd2314451c98e0cb9a37209913" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.753400 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54822ce5f3b2bf8c5856f782cd0169e972fa7ecd2314451c98e0cb9a37209913"} err="failed to get container status \"54822ce5f3b2bf8c5856f782cd0169e972fa7ecd2314451c98e0cb9a37209913\": rpc error: code = NotFound desc = could not find container \"54822ce5f3b2bf8c5856f782cd0169e972fa7ecd2314451c98e0cb9a37209913\": container with ID starting with 54822ce5f3b2bf8c5856f782cd0169e972fa7ecd2314451c98e0cb9a37209913 not found: ID does not exist" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.753441 4931 scope.go:117] "RemoveContainer" containerID="4c039ea56671441919d198cacf9a3a5f441bb04ff71c6660685d90e80193b2a1" Nov 29 00:55:30 crc kubenswrapper[4931]: E1129 00:55:30.754425 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c039ea56671441919d198cacf9a3a5f441bb04ff71c6660685d90e80193b2a1\": container with ID starting with 4c039ea56671441919d198cacf9a3a5f441bb04ff71c6660685d90e80193b2a1 not found: ID does not exist" containerID="4c039ea56671441919d198cacf9a3a5f441bb04ff71c6660685d90e80193b2a1" Nov 29 00:55:30 crc kubenswrapper[4931]: I1129 00:55:30.754467 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c039ea56671441919d198cacf9a3a5f441bb04ff71c6660685d90e80193b2a1"} err="failed to get container status \"4c039ea56671441919d198cacf9a3a5f441bb04ff71c6660685d90e80193b2a1\": rpc error: code = NotFound desc = could not find container \"4c039ea56671441919d198cacf9a3a5f441bb04ff71c6660685d90e80193b2a1\": container with ID starting with 4c039ea56671441919d198cacf9a3a5f441bb04ff71c6660685d90e80193b2a1 not found: ID does not exist" Nov 29 00:55:31 crc kubenswrapper[4931]: I1129 00:55:31.233634 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19cbd316-da53-4f3c-a17a-418c297a74b7" path="/var/lib/kubelet/pods/19cbd316-da53-4f3c-a17a-418c297a74b7/volumes" Nov 29 00:55:39 crc kubenswrapper[4931]: I1129 00:55:39.212852 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:55:39 crc kubenswrapper[4931]: E1129 00:55:39.213993 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:55:52 crc kubenswrapper[4931]: I1129 00:55:52.212664 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:55:52 crc kubenswrapper[4931]: E1129 00:55:52.213791 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:56:07 crc kubenswrapper[4931]: I1129 00:56:07.224173 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:56:07 crc kubenswrapper[4931]: E1129 00:56:07.225037 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:56:19 crc kubenswrapper[4931]: I1129 00:56:19.212597 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:56:19 crc kubenswrapper[4931]: E1129 00:56:19.213458 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.178826 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6n2dq"] Nov 29 00:56:31 crc kubenswrapper[4931]: E1129 00:56:31.179787 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19cbd316-da53-4f3c-a17a-418c297a74b7" containerName="registry-server" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.179802 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="19cbd316-da53-4f3c-a17a-418c297a74b7" containerName="registry-server" Nov 29 00:56:31 crc kubenswrapper[4931]: E1129 00:56:31.180624 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19cbd316-da53-4f3c-a17a-418c297a74b7" containerName="extract-content" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.180644 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="19cbd316-da53-4f3c-a17a-418c297a74b7" containerName="extract-content" Nov 29 00:56:31 crc kubenswrapper[4931]: E1129 00:56:31.180662 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19cbd316-da53-4f3c-a17a-418c297a74b7" containerName="extract-utilities" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.180671 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="19cbd316-da53-4f3c-a17a-418c297a74b7" containerName="extract-utilities" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.180995 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="19cbd316-da53-4f3c-a17a-418c297a74b7" containerName="registry-server" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.182513 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.194310 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6n2dq"] Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.213640 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:56:31 crc kubenswrapper[4931]: E1129 00:56:31.213931 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.327855 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/746ba3db-9a2d-42fb-bdfd-2014e27462d9-utilities\") pod \"certified-operators-6n2dq\" (UID: \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\") " pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.329005 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xltn\" (UniqueName: \"kubernetes.io/projected/746ba3db-9a2d-42fb-bdfd-2014e27462d9-kube-api-access-6xltn\") pod \"certified-operators-6n2dq\" (UID: \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\") " pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.329183 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/746ba3db-9a2d-42fb-bdfd-2014e27462d9-catalog-content\") pod \"certified-operators-6n2dq\" (UID: \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\") " pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.431204 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xltn\" (UniqueName: \"kubernetes.io/projected/746ba3db-9a2d-42fb-bdfd-2014e27462d9-kube-api-access-6xltn\") pod \"certified-operators-6n2dq\" (UID: \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\") " pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.431518 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/746ba3db-9a2d-42fb-bdfd-2014e27462d9-catalog-content\") pod \"certified-operators-6n2dq\" (UID: \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\") " pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.431698 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/746ba3db-9a2d-42fb-bdfd-2014e27462d9-utilities\") pod \"certified-operators-6n2dq\" (UID: \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\") " pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.432485 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/746ba3db-9a2d-42fb-bdfd-2014e27462d9-utilities\") pod \"certified-operators-6n2dq\" (UID: \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\") " pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.433084 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/746ba3db-9a2d-42fb-bdfd-2014e27462d9-catalog-content\") pod \"certified-operators-6n2dq\" (UID: \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\") " pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.472904 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xltn\" (UniqueName: \"kubernetes.io/projected/746ba3db-9a2d-42fb-bdfd-2014e27462d9-kube-api-access-6xltn\") pod \"certified-operators-6n2dq\" (UID: \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\") " pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:31 crc kubenswrapper[4931]: I1129 00:56:31.538514 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:32 crc kubenswrapper[4931]: I1129 00:56:32.046801 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6n2dq"] Nov 29 00:56:32 crc kubenswrapper[4931]: I1129 00:56:32.270461 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6n2dq" event={"ID":"746ba3db-9a2d-42fb-bdfd-2014e27462d9","Type":"ContainerStarted","Data":"feae64310dfab746b7d8e167c563895b438158363bc667ee0c2f1cd7afa8d569"} Nov 29 00:56:33 crc kubenswrapper[4931]: I1129 00:56:33.286161 4931 generic.go:334] "Generic (PLEG): container finished" podID="746ba3db-9a2d-42fb-bdfd-2014e27462d9" containerID="d482392530beffaa45cc490c96c26590b34b35c997f105fdf66021e381e75a5d" exitCode=0 Nov 29 00:56:33 crc kubenswrapper[4931]: I1129 00:56:33.286383 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6n2dq" event={"ID":"746ba3db-9a2d-42fb-bdfd-2014e27462d9","Type":"ContainerDied","Data":"d482392530beffaa45cc490c96c26590b34b35c997f105fdf66021e381e75a5d"} Nov 29 00:56:35 crc kubenswrapper[4931]: I1129 00:56:35.311895 4931 generic.go:334] "Generic (PLEG): container finished" podID="746ba3db-9a2d-42fb-bdfd-2014e27462d9" containerID="bd4d4a774bd37c0970cf1b2aaddc4699c0c91a008a0c47ba9e447b360e96d0ca" exitCode=0 Nov 29 00:56:35 crc kubenswrapper[4931]: I1129 00:56:35.312037 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6n2dq" event={"ID":"746ba3db-9a2d-42fb-bdfd-2014e27462d9","Type":"ContainerDied","Data":"bd4d4a774bd37c0970cf1b2aaddc4699c0c91a008a0c47ba9e447b360e96d0ca"} Nov 29 00:56:36 crc kubenswrapper[4931]: I1129 00:56:36.352080 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6n2dq" event={"ID":"746ba3db-9a2d-42fb-bdfd-2014e27462d9","Type":"ContainerStarted","Data":"defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3"} Nov 29 00:56:36 crc kubenswrapper[4931]: I1129 00:56:36.377333 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6n2dq" podStartSLOduration=2.888516397 podStartE2EDuration="5.377313926s" podCreationTimestamp="2025-11-29 00:56:31 +0000 UTC" firstStartedPulling="2025-11-29 00:56:33.290317492 +0000 UTC m=+3406.452210754" lastFinishedPulling="2025-11-29 00:56:35.779115051 +0000 UTC m=+3408.941008283" observedRunningTime="2025-11-29 00:56:36.368747662 +0000 UTC m=+3409.530640894" watchObservedRunningTime="2025-11-29 00:56:36.377313926 +0000 UTC m=+3409.539207158" Nov 29 00:56:41 crc kubenswrapper[4931]: I1129 00:56:41.539008 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:41 crc kubenswrapper[4931]: I1129 00:56:41.539515 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:41 crc kubenswrapper[4931]: I1129 00:56:41.589570 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:42 crc kubenswrapper[4931]: I1129 00:56:42.496150 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:42 crc kubenswrapper[4931]: I1129 00:56:42.576136 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6n2dq"] Nov 29 00:56:43 crc kubenswrapper[4931]: I1129 00:56:43.212387 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:56:43 crc kubenswrapper[4931]: E1129 00:56:43.213526 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 00:56:44 crc kubenswrapper[4931]: I1129 00:56:44.440087 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6n2dq" podUID="746ba3db-9a2d-42fb-bdfd-2014e27462d9" containerName="registry-server" containerID="cri-o://defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3" gracePeriod=2 Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.057091 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.229022 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/746ba3db-9a2d-42fb-bdfd-2014e27462d9-utilities\") pod \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\" (UID: \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\") " Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.229089 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/746ba3db-9a2d-42fb-bdfd-2014e27462d9-catalog-content\") pod \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\" (UID: \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\") " Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.229238 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xltn\" (UniqueName: \"kubernetes.io/projected/746ba3db-9a2d-42fb-bdfd-2014e27462d9-kube-api-access-6xltn\") pod \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\" (UID: \"746ba3db-9a2d-42fb-bdfd-2014e27462d9\") " Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.231258 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/746ba3db-9a2d-42fb-bdfd-2014e27462d9-utilities" (OuterVolumeSpecName: "utilities") pod "746ba3db-9a2d-42fb-bdfd-2014e27462d9" (UID: "746ba3db-9a2d-42fb-bdfd-2014e27462d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.241754 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/746ba3db-9a2d-42fb-bdfd-2014e27462d9-kube-api-access-6xltn" (OuterVolumeSpecName: "kube-api-access-6xltn") pod "746ba3db-9a2d-42fb-bdfd-2014e27462d9" (UID: "746ba3db-9a2d-42fb-bdfd-2014e27462d9"). InnerVolumeSpecName "kube-api-access-6xltn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.308976 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/746ba3db-9a2d-42fb-bdfd-2014e27462d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "746ba3db-9a2d-42fb-bdfd-2014e27462d9" (UID: "746ba3db-9a2d-42fb-bdfd-2014e27462d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.332382 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/746ba3db-9a2d-42fb-bdfd-2014e27462d9-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.332415 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/746ba3db-9a2d-42fb-bdfd-2014e27462d9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.332428 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xltn\" (UniqueName: \"kubernetes.io/projected/746ba3db-9a2d-42fb-bdfd-2014e27462d9-kube-api-access-6xltn\") on node \"crc\" DevicePath \"\"" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.457914 4931 generic.go:334] "Generic (PLEG): container finished" podID="746ba3db-9a2d-42fb-bdfd-2014e27462d9" containerID="defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3" exitCode=0 Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.457973 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6n2dq" event={"ID":"746ba3db-9a2d-42fb-bdfd-2014e27462d9","Type":"ContainerDied","Data":"defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3"} Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.458014 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6n2dq" event={"ID":"746ba3db-9a2d-42fb-bdfd-2014e27462d9","Type":"ContainerDied","Data":"feae64310dfab746b7d8e167c563895b438158363bc667ee0c2f1cd7afa8d569"} Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.458047 4931 scope.go:117] "RemoveContainer" containerID="defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.458261 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6n2dq" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.497207 4931 scope.go:117] "RemoveContainer" containerID="bd4d4a774bd37c0970cf1b2aaddc4699c0c91a008a0c47ba9e447b360e96d0ca" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.533577 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6n2dq"] Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.551667 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6n2dq"] Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.559057 4931 scope.go:117] "RemoveContainer" containerID="d482392530beffaa45cc490c96c26590b34b35c997f105fdf66021e381e75a5d" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.602437 4931 scope.go:117] "RemoveContainer" containerID="defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3" Nov 29 00:56:45 crc kubenswrapper[4931]: E1129 00:56:45.603174 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3\": container with ID starting with defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3 not found: ID does not exist" containerID="defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.603428 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3"} err="failed to get container status \"defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3\": rpc error: code = NotFound desc = could not find container \"defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3\": container with ID starting with defd2b6d69f4cdf8b17827d0ae6e28099f66bfc1949081dcad85f48554cbeac3 not found: ID does not exist" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.603485 4931 scope.go:117] "RemoveContainer" containerID="bd4d4a774bd37c0970cf1b2aaddc4699c0c91a008a0c47ba9e447b360e96d0ca" Nov 29 00:56:45 crc kubenswrapper[4931]: E1129 00:56:45.604106 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd4d4a774bd37c0970cf1b2aaddc4699c0c91a008a0c47ba9e447b360e96d0ca\": container with ID starting with bd4d4a774bd37c0970cf1b2aaddc4699c0c91a008a0c47ba9e447b360e96d0ca not found: ID does not exist" containerID="bd4d4a774bd37c0970cf1b2aaddc4699c0c91a008a0c47ba9e447b360e96d0ca" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.604152 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd4d4a774bd37c0970cf1b2aaddc4699c0c91a008a0c47ba9e447b360e96d0ca"} err="failed to get container status \"bd4d4a774bd37c0970cf1b2aaddc4699c0c91a008a0c47ba9e447b360e96d0ca\": rpc error: code = NotFound desc = could not find container \"bd4d4a774bd37c0970cf1b2aaddc4699c0c91a008a0c47ba9e447b360e96d0ca\": container with ID starting with bd4d4a774bd37c0970cf1b2aaddc4699c0c91a008a0c47ba9e447b360e96d0ca not found: ID does not exist" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.604180 4931 scope.go:117] "RemoveContainer" containerID="d482392530beffaa45cc490c96c26590b34b35c997f105fdf66021e381e75a5d" Nov 29 00:56:45 crc kubenswrapper[4931]: E1129 00:56:45.605065 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d482392530beffaa45cc490c96c26590b34b35c997f105fdf66021e381e75a5d\": container with ID starting with d482392530beffaa45cc490c96c26590b34b35c997f105fdf66021e381e75a5d not found: ID does not exist" containerID="d482392530beffaa45cc490c96c26590b34b35c997f105fdf66021e381e75a5d" Nov 29 00:56:45 crc kubenswrapper[4931]: I1129 00:56:45.605095 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d482392530beffaa45cc490c96c26590b34b35c997f105fdf66021e381e75a5d"} err="failed to get container status \"d482392530beffaa45cc490c96c26590b34b35c997f105fdf66021e381e75a5d\": rpc error: code = NotFound desc = could not find container \"d482392530beffaa45cc490c96c26590b34b35c997f105fdf66021e381e75a5d\": container with ID starting with d482392530beffaa45cc490c96c26590b34b35c997f105fdf66021e381e75a5d not found: ID does not exist" Nov 29 00:56:47 crc kubenswrapper[4931]: I1129 00:56:47.228353 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="746ba3db-9a2d-42fb-bdfd-2014e27462d9" path="/var/lib/kubelet/pods/746ba3db-9a2d-42fb-bdfd-2014e27462d9/volumes" Nov 29 00:56:54 crc kubenswrapper[4931]: I1129 00:56:54.212498 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 00:56:54 crc kubenswrapper[4931]: I1129 00:56:54.575347 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"089967d9016f33d2d57b978bd4d367473dd4604ca21d5078d10d82bb6e3d56fe"} Nov 29 00:59:23 crc kubenswrapper[4931]: I1129 00:59:23.734651 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:59:23 crc kubenswrapper[4931]: I1129 00:59:23.735327 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.513234 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vjp8k"] Nov 29 00:59:38 crc kubenswrapper[4931]: E1129 00:59:38.514420 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="746ba3db-9a2d-42fb-bdfd-2014e27462d9" containerName="extract-utilities" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.514443 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="746ba3db-9a2d-42fb-bdfd-2014e27462d9" containerName="extract-utilities" Nov 29 00:59:38 crc kubenswrapper[4931]: E1129 00:59:38.514486 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="746ba3db-9a2d-42fb-bdfd-2014e27462d9" containerName="registry-server" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.514500 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="746ba3db-9a2d-42fb-bdfd-2014e27462d9" containerName="registry-server" Nov 29 00:59:38 crc kubenswrapper[4931]: E1129 00:59:38.514525 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="746ba3db-9a2d-42fb-bdfd-2014e27462d9" containerName="extract-content" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.514539 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="746ba3db-9a2d-42fb-bdfd-2014e27462d9" containerName="extract-content" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.514926 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="746ba3db-9a2d-42fb-bdfd-2014e27462d9" containerName="registry-server" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.518382 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.544984 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vjp8k"] Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.614267 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/206b4003-74b7-4913-aa79-4e206bd71ca6-catalog-content\") pod \"redhat-operators-vjp8k\" (UID: \"206b4003-74b7-4913-aa79-4e206bd71ca6\") " pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.614347 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hvtx\" (UniqueName: \"kubernetes.io/projected/206b4003-74b7-4913-aa79-4e206bd71ca6-kube-api-access-7hvtx\") pod \"redhat-operators-vjp8k\" (UID: \"206b4003-74b7-4913-aa79-4e206bd71ca6\") " pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.614389 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/206b4003-74b7-4913-aa79-4e206bd71ca6-utilities\") pod \"redhat-operators-vjp8k\" (UID: \"206b4003-74b7-4913-aa79-4e206bd71ca6\") " pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.717231 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/206b4003-74b7-4913-aa79-4e206bd71ca6-catalog-content\") pod \"redhat-operators-vjp8k\" (UID: \"206b4003-74b7-4913-aa79-4e206bd71ca6\") " pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.717318 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hvtx\" (UniqueName: \"kubernetes.io/projected/206b4003-74b7-4913-aa79-4e206bd71ca6-kube-api-access-7hvtx\") pod \"redhat-operators-vjp8k\" (UID: \"206b4003-74b7-4913-aa79-4e206bd71ca6\") " pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.717357 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/206b4003-74b7-4913-aa79-4e206bd71ca6-utilities\") pod \"redhat-operators-vjp8k\" (UID: \"206b4003-74b7-4913-aa79-4e206bd71ca6\") " pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.718018 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/206b4003-74b7-4913-aa79-4e206bd71ca6-utilities\") pod \"redhat-operators-vjp8k\" (UID: \"206b4003-74b7-4913-aa79-4e206bd71ca6\") " pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.718664 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/206b4003-74b7-4913-aa79-4e206bd71ca6-catalog-content\") pod \"redhat-operators-vjp8k\" (UID: \"206b4003-74b7-4913-aa79-4e206bd71ca6\") " pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.741497 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hvtx\" (UniqueName: \"kubernetes.io/projected/206b4003-74b7-4913-aa79-4e206bd71ca6-kube-api-access-7hvtx\") pod \"redhat-operators-vjp8k\" (UID: \"206b4003-74b7-4913-aa79-4e206bd71ca6\") " pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:38 crc kubenswrapper[4931]: I1129 00:59:38.853675 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:39 crc kubenswrapper[4931]: I1129 00:59:39.332707 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vjp8k"] Nov 29 00:59:39 crc kubenswrapper[4931]: I1129 00:59:39.531338 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vjp8k" event={"ID":"206b4003-74b7-4913-aa79-4e206bd71ca6","Type":"ContainerStarted","Data":"03a853c843a8f142de57ae806fe060636854f31622fe8f38adb210f5e13785e3"} Nov 29 00:59:40 crc kubenswrapper[4931]: I1129 00:59:40.544584 4931 generic.go:334] "Generic (PLEG): container finished" podID="206b4003-74b7-4913-aa79-4e206bd71ca6" containerID="2385b5a3ff78265c52990105c004d74d71d5c9139e4677074e2339473b48f7c7" exitCode=0 Nov 29 00:59:40 crc kubenswrapper[4931]: I1129 00:59:40.544697 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vjp8k" event={"ID":"206b4003-74b7-4913-aa79-4e206bd71ca6","Type":"ContainerDied","Data":"2385b5a3ff78265c52990105c004d74d71d5c9139e4677074e2339473b48f7c7"} Nov 29 00:59:40 crc kubenswrapper[4931]: I1129 00:59:40.549442 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 00:59:42 crc kubenswrapper[4931]: I1129 00:59:42.566188 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vjp8k" event={"ID":"206b4003-74b7-4913-aa79-4e206bd71ca6","Type":"ContainerStarted","Data":"3a0dc73f6cb57815953adcd8f577aed196a7395d25b13471b3cd5fa56e48b31e"} Nov 29 00:59:44 crc kubenswrapper[4931]: I1129 00:59:44.599848 4931 generic.go:334] "Generic (PLEG): container finished" podID="206b4003-74b7-4913-aa79-4e206bd71ca6" containerID="3a0dc73f6cb57815953adcd8f577aed196a7395d25b13471b3cd5fa56e48b31e" exitCode=0 Nov 29 00:59:44 crc kubenswrapper[4931]: I1129 00:59:44.599951 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vjp8k" event={"ID":"206b4003-74b7-4913-aa79-4e206bd71ca6","Type":"ContainerDied","Data":"3a0dc73f6cb57815953adcd8f577aed196a7395d25b13471b3cd5fa56e48b31e"} Nov 29 00:59:45 crc kubenswrapper[4931]: I1129 00:59:45.636393 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vjp8k" event={"ID":"206b4003-74b7-4913-aa79-4e206bd71ca6","Type":"ContainerStarted","Data":"23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2"} Nov 29 00:59:45 crc kubenswrapper[4931]: I1129 00:59:45.672082 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vjp8k" podStartSLOduration=3.157698516 podStartE2EDuration="7.672055101s" podCreationTimestamp="2025-11-29 00:59:38 +0000 UTC" firstStartedPulling="2025-11-29 00:59:40.54915308 +0000 UTC m=+3593.711046312" lastFinishedPulling="2025-11-29 00:59:45.063509625 +0000 UTC m=+3598.225402897" observedRunningTime="2025-11-29 00:59:45.66173809 +0000 UTC m=+3598.823631362" watchObservedRunningTime="2025-11-29 00:59:45.672055101 +0000 UTC m=+3598.833948363" Nov 29 00:59:48 crc kubenswrapper[4931]: I1129 00:59:48.854868 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:48 crc kubenswrapper[4931]: I1129 00:59:48.857070 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:49 crc kubenswrapper[4931]: I1129 00:59:49.904570 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vjp8k" podUID="206b4003-74b7-4913-aa79-4e206bd71ca6" containerName="registry-server" probeResult="failure" output=< Nov 29 00:59:49 crc kubenswrapper[4931]: timeout: failed to connect service ":50051" within 1s Nov 29 00:59:49 crc kubenswrapper[4931]: > Nov 29 00:59:52 crc kubenswrapper[4931]: I1129 00:59:52.887202 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5z5pr"] Nov 29 00:59:52 crc kubenswrapper[4931]: I1129 00:59:52.892121 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5z5pr" Nov 29 00:59:52 crc kubenswrapper[4931]: I1129 00:59:52.912907 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5z5pr"] Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.056036 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30f61348-5315-4a92-9a3e-dc3f271da38e-catalog-content\") pod \"community-operators-5z5pr\" (UID: \"30f61348-5315-4a92-9a3e-dc3f271da38e\") " pod="openshift-marketplace/community-operators-5z5pr" Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.056128 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30f61348-5315-4a92-9a3e-dc3f271da38e-utilities\") pod \"community-operators-5z5pr\" (UID: \"30f61348-5315-4a92-9a3e-dc3f271da38e\") " pod="openshift-marketplace/community-operators-5z5pr" Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.056284 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vk8q\" (UniqueName: \"kubernetes.io/projected/30f61348-5315-4a92-9a3e-dc3f271da38e-kube-api-access-8vk8q\") pod \"community-operators-5z5pr\" (UID: \"30f61348-5315-4a92-9a3e-dc3f271da38e\") " pod="openshift-marketplace/community-operators-5z5pr" Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.158102 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vk8q\" (UniqueName: \"kubernetes.io/projected/30f61348-5315-4a92-9a3e-dc3f271da38e-kube-api-access-8vk8q\") pod \"community-operators-5z5pr\" (UID: \"30f61348-5315-4a92-9a3e-dc3f271da38e\") " pod="openshift-marketplace/community-operators-5z5pr" Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.158678 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30f61348-5315-4a92-9a3e-dc3f271da38e-catalog-content\") pod \"community-operators-5z5pr\" (UID: \"30f61348-5315-4a92-9a3e-dc3f271da38e\") " pod="openshift-marketplace/community-operators-5z5pr" Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.159250 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30f61348-5315-4a92-9a3e-dc3f271da38e-catalog-content\") pod \"community-operators-5z5pr\" (UID: \"30f61348-5315-4a92-9a3e-dc3f271da38e\") " pod="openshift-marketplace/community-operators-5z5pr" Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.159292 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30f61348-5315-4a92-9a3e-dc3f271da38e-utilities\") pod \"community-operators-5z5pr\" (UID: \"30f61348-5315-4a92-9a3e-dc3f271da38e\") " pod="openshift-marketplace/community-operators-5z5pr" Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.159545 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30f61348-5315-4a92-9a3e-dc3f271da38e-utilities\") pod \"community-operators-5z5pr\" (UID: \"30f61348-5315-4a92-9a3e-dc3f271da38e\") " pod="openshift-marketplace/community-operators-5z5pr" Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.188849 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vk8q\" (UniqueName: \"kubernetes.io/projected/30f61348-5315-4a92-9a3e-dc3f271da38e-kube-api-access-8vk8q\") pod \"community-operators-5z5pr\" (UID: \"30f61348-5315-4a92-9a3e-dc3f271da38e\") " pod="openshift-marketplace/community-operators-5z5pr" Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.238211 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5z5pr" Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.734722 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.734786 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 00:59:53 crc kubenswrapper[4931]: I1129 00:59:53.753966 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5z5pr"] Nov 29 00:59:54 crc kubenswrapper[4931]: I1129 00:59:54.737728 4931 generic.go:334] "Generic (PLEG): container finished" podID="30f61348-5315-4a92-9a3e-dc3f271da38e" containerID="cfb1510b7d4fcd75b57b3bb0e15d596c7828fb795316a2c1fd0953efcbe018cf" exitCode=0 Nov 29 00:59:54 crc kubenswrapper[4931]: I1129 00:59:54.737788 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5z5pr" event={"ID":"30f61348-5315-4a92-9a3e-dc3f271da38e","Type":"ContainerDied","Data":"cfb1510b7d4fcd75b57b3bb0e15d596c7828fb795316a2c1fd0953efcbe018cf"} Nov 29 00:59:54 crc kubenswrapper[4931]: I1129 00:59:54.738263 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5z5pr" event={"ID":"30f61348-5315-4a92-9a3e-dc3f271da38e","Type":"ContainerStarted","Data":"0451adb22b898e5e503a3734c9bec569c7a389f66463201dccef8dab20ace463"} Nov 29 00:59:55 crc kubenswrapper[4931]: I1129 00:59:55.752385 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5z5pr" event={"ID":"30f61348-5315-4a92-9a3e-dc3f271da38e","Type":"ContainerStarted","Data":"3f45158f17a90d9ef50223deb7f26b923b93af25e9d8844d755c6af5b4fc9214"} Nov 29 00:59:56 crc kubenswrapper[4931]: I1129 00:59:56.772687 4931 generic.go:334] "Generic (PLEG): container finished" podID="30f61348-5315-4a92-9a3e-dc3f271da38e" containerID="3f45158f17a90d9ef50223deb7f26b923b93af25e9d8844d755c6af5b4fc9214" exitCode=0 Nov 29 00:59:56 crc kubenswrapper[4931]: I1129 00:59:56.773291 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5z5pr" event={"ID":"30f61348-5315-4a92-9a3e-dc3f271da38e","Type":"ContainerDied","Data":"3f45158f17a90d9ef50223deb7f26b923b93af25e9d8844d755c6af5b4fc9214"} Nov 29 00:59:57 crc kubenswrapper[4931]: I1129 00:59:57.783987 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5z5pr" event={"ID":"30f61348-5315-4a92-9a3e-dc3f271da38e","Type":"ContainerStarted","Data":"7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc"} Nov 29 00:59:57 crc kubenswrapper[4931]: I1129 00:59:57.815192 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5z5pr" podStartSLOduration=3.174293793 podStartE2EDuration="5.815168726s" podCreationTimestamp="2025-11-29 00:59:52 +0000 UTC" firstStartedPulling="2025-11-29 00:59:54.739889704 +0000 UTC m=+3607.901782976" lastFinishedPulling="2025-11-29 00:59:57.380764657 +0000 UTC m=+3610.542657909" observedRunningTime="2025-11-29 00:59:57.80539522 +0000 UTC m=+3610.967288462" watchObservedRunningTime="2025-11-29 00:59:57.815168726 +0000 UTC m=+3610.977061958" Nov 29 00:59:58 crc kubenswrapper[4931]: I1129 00:59:58.921572 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 00:59:58 crc kubenswrapper[4931]: I1129 00:59:58.991906 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.166089 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v"] Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.167691 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.170299 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.170338 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.177821 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v"] Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.248101 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vjp8k"] Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.314594 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-secret-volume\") pod \"collect-profiles-29406300-lrs4v\" (UID: \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.314747 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmbql\" (UniqueName: \"kubernetes.io/projected/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-kube-api-access-xmbql\") pod \"collect-profiles-29406300-lrs4v\" (UID: \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.315151 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-config-volume\") pod \"collect-profiles-29406300-lrs4v\" (UID: \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.416844 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-secret-volume\") pod \"collect-profiles-29406300-lrs4v\" (UID: \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.417251 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmbql\" (UniqueName: \"kubernetes.io/projected/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-kube-api-access-xmbql\") pod \"collect-profiles-29406300-lrs4v\" (UID: \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.417377 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-config-volume\") pod \"collect-profiles-29406300-lrs4v\" (UID: \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.418377 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-config-volume\") pod \"collect-profiles-29406300-lrs4v\" (UID: \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.425242 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-secret-volume\") pod \"collect-profiles-29406300-lrs4v\" (UID: \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.435222 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmbql\" (UniqueName: \"kubernetes.io/projected/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-kube-api-access-xmbql\") pod \"collect-profiles-29406300-lrs4v\" (UID: \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.534703 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:00 crc kubenswrapper[4931]: I1129 01:00:00.818220 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vjp8k" podUID="206b4003-74b7-4913-aa79-4e206bd71ca6" containerName="registry-server" containerID="cri-o://23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2" gracePeriod=2 Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.030788 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v"] Nov 29 01:00:01 crc kubenswrapper[4931]: W1129 01:00:01.069177 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d94a4c2_728a_4f7f_ad4a_1d7a0ae64da5.slice/crio-a444d0ed361463e8480aff1933fbee013e69a7803af6eecbd40d51164fc22205 WatchSource:0}: Error finding container a444d0ed361463e8480aff1933fbee013e69a7803af6eecbd40d51164fc22205: Status 404 returned error can't find the container with id a444d0ed361463e8480aff1933fbee013e69a7803af6eecbd40d51164fc22205 Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.368221 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.546172 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/206b4003-74b7-4913-aa79-4e206bd71ca6-utilities\") pod \"206b4003-74b7-4913-aa79-4e206bd71ca6\" (UID: \"206b4003-74b7-4913-aa79-4e206bd71ca6\") " Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.546413 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/206b4003-74b7-4913-aa79-4e206bd71ca6-catalog-content\") pod \"206b4003-74b7-4913-aa79-4e206bd71ca6\" (UID: \"206b4003-74b7-4913-aa79-4e206bd71ca6\") " Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.546539 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hvtx\" (UniqueName: \"kubernetes.io/projected/206b4003-74b7-4913-aa79-4e206bd71ca6-kube-api-access-7hvtx\") pod \"206b4003-74b7-4913-aa79-4e206bd71ca6\" (UID: \"206b4003-74b7-4913-aa79-4e206bd71ca6\") " Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.547110 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/206b4003-74b7-4913-aa79-4e206bd71ca6-utilities" (OuterVolumeSpecName: "utilities") pod "206b4003-74b7-4913-aa79-4e206bd71ca6" (UID: "206b4003-74b7-4913-aa79-4e206bd71ca6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.547587 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/206b4003-74b7-4913-aa79-4e206bd71ca6-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.552014 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/206b4003-74b7-4913-aa79-4e206bd71ca6-kube-api-access-7hvtx" (OuterVolumeSpecName: "kube-api-access-7hvtx") pod "206b4003-74b7-4913-aa79-4e206bd71ca6" (UID: "206b4003-74b7-4913-aa79-4e206bd71ca6"). InnerVolumeSpecName "kube-api-access-7hvtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.649493 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hvtx\" (UniqueName: \"kubernetes.io/projected/206b4003-74b7-4913-aa79-4e206bd71ca6-kube-api-access-7hvtx\") on node \"crc\" DevicePath \"\"" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.659473 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/206b4003-74b7-4913-aa79-4e206bd71ca6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "206b4003-74b7-4913-aa79-4e206bd71ca6" (UID: "206b4003-74b7-4913-aa79-4e206bd71ca6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.750985 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/206b4003-74b7-4913-aa79-4e206bd71ca6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.832627 4931 generic.go:334] "Generic (PLEG): container finished" podID="206b4003-74b7-4913-aa79-4e206bd71ca6" containerID="23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2" exitCode=0 Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.832735 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vjp8k" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.832774 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vjp8k" event={"ID":"206b4003-74b7-4913-aa79-4e206bd71ca6","Type":"ContainerDied","Data":"23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2"} Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.833232 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vjp8k" event={"ID":"206b4003-74b7-4913-aa79-4e206bd71ca6","Type":"ContainerDied","Data":"03a853c843a8f142de57ae806fe060636854f31622fe8f38adb210f5e13785e3"} Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.833278 4931 scope.go:117] "RemoveContainer" containerID="23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.837005 4931 generic.go:334] "Generic (PLEG): container finished" podID="1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5" containerID="3ba2892b85da50601ee4e2e13b5837e84b4cb72139262b77482f14af01f5d9f9" exitCode=0 Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.837078 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" event={"ID":"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5","Type":"ContainerDied","Data":"3ba2892b85da50601ee4e2e13b5837e84b4cb72139262b77482f14af01f5d9f9"} Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.837134 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" event={"ID":"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5","Type":"ContainerStarted","Data":"a444d0ed361463e8480aff1933fbee013e69a7803af6eecbd40d51164fc22205"} Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.867691 4931 scope.go:117] "RemoveContainer" containerID="3a0dc73f6cb57815953adcd8f577aed196a7395d25b13471b3cd5fa56e48b31e" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.887173 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vjp8k"] Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.898721 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vjp8k"] Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.910766 4931 scope.go:117] "RemoveContainer" containerID="2385b5a3ff78265c52990105c004d74d71d5c9139e4677074e2339473b48f7c7" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.944395 4931 scope.go:117] "RemoveContainer" containerID="23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2" Nov 29 01:00:01 crc kubenswrapper[4931]: E1129 01:00:01.949067 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2\": container with ID starting with 23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2 not found: ID does not exist" containerID="23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.949276 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2"} err="failed to get container status \"23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2\": rpc error: code = NotFound desc = could not find container \"23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2\": container with ID starting with 23e3088fe334e501de57ec817b9b03224000693bfc135661c68329189a1781f2 not found: ID does not exist" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.949418 4931 scope.go:117] "RemoveContainer" containerID="3a0dc73f6cb57815953adcd8f577aed196a7395d25b13471b3cd5fa56e48b31e" Nov 29 01:00:01 crc kubenswrapper[4931]: E1129 01:00:01.949865 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a0dc73f6cb57815953adcd8f577aed196a7395d25b13471b3cd5fa56e48b31e\": container with ID starting with 3a0dc73f6cb57815953adcd8f577aed196a7395d25b13471b3cd5fa56e48b31e not found: ID does not exist" containerID="3a0dc73f6cb57815953adcd8f577aed196a7395d25b13471b3cd5fa56e48b31e" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.949996 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a0dc73f6cb57815953adcd8f577aed196a7395d25b13471b3cd5fa56e48b31e"} err="failed to get container status \"3a0dc73f6cb57815953adcd8f577aed196a7395d25b13471b3cd5fa56e48b31e\": rpc error: code = NotFound desc = could not find container \"3a0dc73f6cb57815953adcd8f577aed196a7395d25b13471b3cd5fa56e48b31e\": container with ID starting with 3a0dc73f6cb57815953adcd8f577aed196a7395d25b13471b3cd5fa56e48b31e not found: ID does not exist" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.950111 4931 scope.go:117] "RemoveContainer" containerID="2385b5a3ff78265c52990105c004d74d71d5c9139e4677074e2339473b48f7c7" Nov 29 01:00:01 crc kubenswrapper[4931]: E1129 01:00:01.950529 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2385b5a3ff78265c52990105c004d74d71d5c9139e4677074e2339473b48f7c7\": container with ID starting with 2385b5a3ff78265c52990105c004d74d71d5c9139e4677074e2339473b48f7c7 not found: ID does not exist" containerID="2385b5a3ff78265c52990105c004d74d71d5c9139e4677074e2339473b48f7c7" Nov 29 01:00:01 crc kubenswrapper[4931]: I1129 01:00:01.950634 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2385b5a3ff78265c52990105c004d74d71d5c9139e4677074e2339473b48f7c7"} err="failed to get container status \"2385b5a3ff78265c52990105c004d74d71d5c9139e4677074e2339473b48f7c7\": rpc error: code = NotFound desc = could not find container \"2385b5a3ff78265c52990105c004d74d71d5c9139e4677074e2339473b48f7c7\": container with ID starting with 2385b5a3ff78265c52990105c004d74d71d5c9139e4677074e2339473b48f7c7 not found: ID does not exist" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.226118 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="206b4003-74b7-4913-aa79-4e206bd71ca6" path="/var/lib/kubelet/pods/206b4003-74b7-4913-aa79-4e206bd71ca6/volumes" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.239424 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5z5pr" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.239470 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5z5pr" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.272445 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.327752 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5z5pr" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.382238 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-secret-volume\") pod \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\" (UID: \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\") " Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.382361 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-config-volume\") pod \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\" (UID: \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\") " Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.382511 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmbql\" (UniqueName: \"kubernetes.io/projected/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-kube-api-access-xmbql\") pod \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\" (UID: \"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5\") " Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.383099 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-config-volume" (OuterVolumeSpecName: "config-volume") pod "1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5" (UID: "1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.388778 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-kube-api-access-xmbql" (OuterVolumeSpecName: "kube-api-access-xmbql") pod "1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5" (UID: "1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5"). InnerVolumeSpecName "kube-api-access-xmbql". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.389122 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5" (UID: "1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.484579 4931 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.484618 4931 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.484630 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmbql\" (UniqueName: \"kubernetes.io/projected/1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5-kube-api-access-xmbql\") on node \"crc\" DevicePath \"\"" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.864515 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" event={"ID":"1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5","Type":"ContainerDied","Data":"a444d0ed361463e8480aff1933fbee013e69a7803af6eecbd40d51164fc22205"} Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.864554 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406300-lrs4v" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.864868 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a444d0ed361463e8480aff1933fbee013e69a7803af6eecbd40d51164fc22205" Nov 29 01:00:03 crc kubenswrapper[4931]: I1129 01:00:03.933868 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5z5pr" Nov 29 01:00:04 crc kubenswrapper[4931]: I1129 01:00:04.367008 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj"] Nov 29 01:00:04 crc kubenswrapper[4931]: I1129 01:00:04.374993 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406255-pz6sj"] Nov 29 01:00:04 crc kubenswrapper[4931]: I1129 01:00:04.649832 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5z5pr"] Nov 29 01:00:05 crc kubenswrapper[4931]: I1129 01:00:05.229600 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a" path="/var/lib/kubelet/pods/0f0cfdd5-2bcd-4709-bef6-82a3a1939e2a/volumes" Nov 29 01:00:05 crc kubenswrapper[4931]: I1129 01:00:05.886084 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5z5pr" podUID="30f61348-5315-4a92-9a3e-dc3f271da38e" containerName="registry-server" containerID="cri-o://7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc" gracePeriod=2 Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.368493 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5z5pr" Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.448267 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30f61348-5315-4a92-9a3e-dc3f271da38e-catalog-content\") pod \"30f61348-5315-4a92-9a3e-dc3f271da38e\" (UID: \"30f61348-5315-4a92-9a3e-dc3f271da38e\") " Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.458083 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30f61348-5315-4a92-9a3e-dc3f271da38e-utilities\") pod \"30f61348-5315-4a92-9a3e-dc3f271da38e\" (UID: \"30f61348-5315-4a92-9a3e-dc3f271da38e\") " Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.458238 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vk8q\" (UniqueName: \"kubernetes.io/projected/30f61348-5315-4a92-9a3e-dc3f271da38e-kube-api-access-8vk8q\") pod \"30f61348-5315-4a92-9a3e-dc3f271da38e\" (UID: \"30f61348-5315-4a92-9a3e-dc3f271da38e\") " Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.459335 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30f61348-5315-4a92-9a3e-dc3f271da38e-utilities" (OuterVolumeSpecName: "utilities") pod "30f61348-5315-4a92-9a3e-dc3f271da38e" (UID: "30f61348-5315-4a92-9a3e-dc3f271da38e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.468607 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30f61348-5315-4a92-9a3e-dc3f271da38e-kube-api-access-8vk8q" (OuterVolumeSpecName: "kube-api-access-8vk8q") pod "30f61348-5315-4a92-9a3e-dc3f271da38e" (UID: "30f61348-5315-4a92-9a3e-dc3f271da38e"). InnerVolumeSpecName "kube-api-access-8vk8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.510652 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30f61348-5315-4a92-9a3e-dc3f271da38e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30f61348-5315-4a92-9a3e-dc3f271da38e" (UID: "30f61348-5315-4a92-9a3e-dc3f271da38e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.560904 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30f61348-5315-4a92-9a3e-dc3f271da38e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.560955 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30f61348-5315-4a92-9a3e-dc3f271da38e-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.560979 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vk8q\" (UniqueName: \"kubernetes.io/projected/30f61348-5315-4a92-9a3e-dc3f271da38e-kube-api-access-8vk8q\") on node \"crc\" DevicePath \"\"" Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.910155 4931 generic.go:334] "Generic (PLEG): container finished" podID="30f61348-5315-4a92-9a3e-dc3f271da38e" containerID="7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc" exitCode=0 Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.910195 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5z5pr" event={"ID":"30f61348-5315-4a92-9a3e-dc3f271da38e","Type":"ContainerDied","Data":"7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc"} Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.910222 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5z5pr" event={"ID":"30f61348-5315-4a92-9a3e-dc3f271da38e","Type":"ContainerDied","Data":"0451adb22b898e5e503a3734c9bec569c7a389f66463201dccef8dab20ace463"} Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.910238 4931 scope.go:117] "RemoveContainer" containerID="7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc" Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.910382 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5z5pr" Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.950230 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5z5pr"] Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.970136 4931 scope.go:117] "RemoveContainer" containerID="3f45158f17a90d9ef50223deb7f26b923b93af25e9d8844d755c6af5b4fc9214" Nov 29 01:00:06 crc kubenswrapper[4931]: I1129 01:00:06.970893 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5z5pr"] Nov 29 01:00:07 crc kubenswrapper[4931]: I1129 01:00:07.015710 4931 scope.go:117] "RemoveContainer" containerID="cfb1510b7d4fcd75b57b3bb0e15d596c7828fb795316a2c1fd0953efcbe018cf" Nov 29 01:00:07 crc kubenswrapper[4931]: I1129 01:00:07.053090 4931 scope.go:117] "RemoveContainer" containerID="7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc" Nov 29 01:00:07 crc kubenswrapper[4931]: E1129 01:00:07.053434 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc\": container with ID starting with 7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc not found: ID does not exist" containerID="7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc" Nov 29 01:00:07 crc kubenswrapper[4931]: I1129 01:00:07.053464 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc"} err="failed to get container status \"7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc\": rpc error: code = NotFound desc = could not find container \"7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc\": container with ID starting with 7595f12303dd381f09fb033db800efea81da9d8e5b732d858743a92cc0a8bccc not found: ID does not exist" Nov 29 01:00:07 crc kubenswrapper[4931]: I1129 01:00:07.053484 4931 scope.go:117] "RemoveContainer" containerID="3f45158f17a90d9ef50223deb7f26b923b93af25e9d8844d755c6af5b4fc9214" Nov 29 01:00:07 crc kubenswrapper[4931]: E1129 01:00:07.053987 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f45158f17a90d9ef50223deb7f26b923b93af25e9d8844d755c6af5b4fc9214\": container with ID starting with 3f45158f17a90d9ef50223deb7f26b923b93af25e9d8844d755c6af5b4fc9214 not found: ID does not exist" containerID="3f45158f17a90d9ef50223deb7f26b923b93af25e9d8844d755c6af5b4fc9214" Nov 29 01:00:07 crc kubenswrapper[4931]: I1129 01:00:07.054027 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f45158f17a90d9ef50223deb7f26b923b93af25e9d8844d755c6af5b4fc9214"} err="failed to get container status \"3f45158f17a90d9ef50223deb7f26b923b93af25e9d8844d755c6af5b4fc9214\": rpc error: code = NotFound desc = could not find container \"3f45158f17a90d9ef50223deb7f26b923b93af25e9d8844d755c6af5b4fc9214\": container with ID starting with 3f45158f17a90d9ef50223deb7f26b923b93af25e9d8844d755c6af5b4fc9214 not found: ID does not exist" Nov 29 01:00:07 crc kubenswrapper[4931]: I1129 01:00:07.054053 4931 scope.go:117] "RemoveContainer" containerID="cfb1510b7d4fcd75b57b3bb0e15d596c7828fb795316a2c1fd0953efcbe018cf" Nov 29 01:00:07 crc kubenswrapper[4931]: E1129 01:00:07.054336 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfb1510b7d4fcd75b57b3bb0e15d596c7828fb795316a2c1fd0953efcbe018cf\": container with ID starting with cfb1510b7d4fcd75b57b3bb0e15d596c7828fb795316a2c1fd0953efcbe018cf not found: ID does not exist" containerID="cfb1510b7d4fcd75b57b3bb0e15d596c7828fb795316a2c1fd0953efcbe018cf" Nov 29 01:00:07 crc kubenswrapper[4931]: I1129 01:00:07.054360 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfb1510b7d4fcd75b57b3bb0e15d596c7828fb795316a2c1fd0953efcbe018cf"} err="failed to get container status \"cfb1510b7d4fcd75b57b3bb0e15d596c7828fb795316a2c1fd0953efcbe018cf\": rpc error: code = NotFound desc = could not find container \"cfb1510b7d4fcd75b57b3bb0e15d596c7828fb795316a2c1fd0953efcbe018cf\": container with ID starting with cfb1510b7d4fcd75b57b3bb0e15d596c7828fb795316a2c1fd0953efcbe018cf not found: ID does not exist" Nov 29 01:00:07 crc kubenswrapper[4931]: I1129 01:00:07.229869 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30f61348-5315-4a92-9a3e-dc3f271da38e" path="/var/lib/kubelet/pods/30f61348-5315-4a92-9a3e-dc3f271da38e/volumes" Nov 29 01:00:17 crc kubenswrapper[4931]: I1129 01:00:17.303716 4931 scope.go:117] "RemoveContainer" containerID="aa90dfd80e94e8ba9a47640a271f681b7a918f24890f3002c66d610ab88b01f2" Nov 29 01:00:23 crc kubenswrapper[4931]: I1129 01:00:23.734707 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 01:00:23 crc kubenswrapper[4931]: I1129 01:00:23.735369 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 01:00:23 crc kubenswrapper[4931]: I1129 01:00:23.735440 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 01:00:23 crc kubenswrapper[4931]: I1129 01:00:23.736484 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"089967d9016f33d2d57b978bd4d367473dd4604ca21d5078d10d82bb6e3d56fe"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 01:00:23 crc kubenswrapper[4931]: I1129 01:00:23.736574 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://089967d9016f33d2d57b978bd4d367473dd4604ca21d5078d10d82bb6e3d56fe" gracePeriod=600 Nov 29 01:00:24 crc kubenswrapper[4931]: I1129 01:00:24.111852 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="089967d9016f33d2d57b978bd4d367473dd4604ca21d5078d10d82bb6e3d56fe" exitCode=0 Nov 29 01:00:24 crc kubenswrapper[4931]: I1129 01:00:24.111950 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"089967d9016f33d2d57b978bd4d367473dd4604ca21d5078d10d82bb6e3d56fe"} Nov 29 01:00:24 crc kubenswrapper[4931]: I1129 01:00:24.112220 4931 scope.go:117] "RemoveContainer" containerID="f0f4393cb312f95d2e03296769307475cba3db003b77ab78abbfa955d301a251" Nov 29 01:00:25 crc kubenswrapper[4931]: I1129 01:00:25.128759 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094"} Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.203919 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29406301-hhnfj"] Nov 29 01:01:00 crc kubenswrapper[4931]: E1129 01:01:00.206576 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206b4003-74b7-4913-aa79-4e206bd71ca6" containerName="registry-server" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.206736 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="206b4003-74b7-4913-aa79-4e206bd71ca6" containerName="registry-server" Nov 29 01:01:00 crc kubenswrapper[4931]: E1129 01:01:00.206889 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206b4003-74b7-4913-aa79-4e206bd71ca6" containerName="extract-utilities" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.207021 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="206b4003-74b7-4913-aa79-4e206bd71ca6" containerName="extract-utilities" Nov 29 01:01:00 crc kubenswrapper[4931]: E1129 01:01:00.207169 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30f61348-5315-4a92-9a3e-dc3f271da38e" containerName="extract-content" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.207303 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="30f61348-5315-4a92-9a3e-dc3f271da38e" containerName="extract-content" Nov 29 01:01:00 crc kubenswrapper[4931]: E1129 01:01:00.207437 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5" containerName="collect-profiles" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.207547 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5" containerName="collect-profiles" Nov 29 01:01:00 crc kubenswrapper[4931]: E1129 01:01:00.207701 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206b4003-74b7-4913-aa79-4e206bd71ca6" containerName="extract-content" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.207849 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="206b4003-74b7-4913-aa79-4e206bd71ca6" containerName="extract-content" Nov 29 01:01:00 crc kubenswrapper[4931]: E1129 01:01:00.208004 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30f61348-5315-4a92-9a3e-dc3f271da38e" containerName="extract-utilities" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.208131 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="30f61348-5315-4a92-9a3e-dc3f271da38e" containerName="extract-utilities" Nov 29 01:01:00 crc kubenswrapper[4931]: E1129 01:01:00.208284 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30f61348-5315-4a92-9a3e-dc3f271da38e" containerName="registry-server" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.208403 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="30f61348-5315-4a92-9a3e-dc3f271da38e" containerName="registry-server" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.208868 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="30f61348-5315-4a92-9a3e-dc3f271da38e" containerName="registry-server" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.209036 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d94a4c2-728a-4f7f-ad4a-1d7a0ae64da5" containerName="collect-profiles" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.209192 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="206b4003-74b7-4913-aa79-4e206bd71ca6" containerName="registry-server" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.210337 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.219994 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29406301-hhnfj"] Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.310957 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-config-data\") pod \"keystone-cron-29406301-hhnfj\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.311175 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-fernet-keys\") pod \"keystone-cron-29406301-hhnfj\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.311348 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-combined-ca-bundle\") pod \"keystone-cron-29406301-hhnfj\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.311793 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74nt6\" (UniqueName: \"kubernetes.io/projected/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-kube-api-access-74nt6\") pod \"keystone-cron-29406301-hhnfj\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.413972 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74nt6\" (UniqueName: \"kubernetes.io/projected/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-kube-api-access-74nt6\") pod \"keystone-cron-29406301-hhnfj\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.414160 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-config-data\") pod \"keystone-cron-29406301-hhnfj\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.414223 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-fernet-keys\") pod \"keystone-cron-29406301-hhnfj\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.414294 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-combined-ca-bundle\") pod \"keystone-cron-29406301-hhnfj\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.424189 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-combined-ca-bundle\") pod \"keystone-cron-29406301-hhnfj\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.424891 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-fernet-keys\") pod \"keystone-cron-29406301-hhnfj\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.425334 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-config-data\") pod \"keystone-cron-29406301-hhnfj\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.437412 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74nt6\" (UniqueName: \"kubernetes.io/projected/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-kube-api-access-74nt6\") pod \"keystone-cron-29406301-hhnfj\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:00 crc kubenswrapper[4931]: I1129 01:01:00.547300 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:01 crc kubenswrapper[4931]: I1129 01:01:01.015459 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29406301-hhnfj"] Nov 29 01:01:01 crc kubenswrapper[4931]: I1129 01:01:01.532559 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29406301-hhnfj" event={"ID":"55460ba8-a9d1-4ef2-9bcf-e08d863698fa","Type":"ContainerStarted","Data":"17757fc7390dc6deda9faf0a73c46419b174789f98c6fe90241f7d1f48e7047f"} Nov 29 01:01:01 crc kubenswrapper[4931]: I1129 01:01:01.532609 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29406301-hhnfj" event={"ID":"55460ba8-a9d1-4ef2-9bcf-e08d863698fa","Type":"ContainerStarted","Data":"f2c65ce9c32b900b3000c2c4317d8357fbd48830bb550f2aa0a5ec62f7bde3ba"} Nov 29 01:01:01 crc kubenswrapper[4931]: I1129 01:01:01.552047 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29406301-hhnfj" podStartSLOduration=1.552026578 podStartE2EDuration="1.552026578s" podCreationTimestamp="2025-11-29 01:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 01:01:01.547010646 +0000 UTC m=+3674.708903888" watchObservedRunningTime="2025-11-29 01:01:01.552026578 +0000 UTC m=+3674.713919810" Nov 29 01:01:03 crc kubenswrapper[4931]: I1129 01:01:03.557477 4931 generic.go:334] "Generic (PLEG): container finished" podID="55460ba8-a9d1-4ef2-9bcf-e08d863698fa" containerID="17757fc7390dc6deda9faf0a73c46419b174789f98c6fe90241f7d1f48e7047f" exitCode=0 Nov 29 01:01:03 crc kubenswrapper[4931]: I1129 01:01:03.557599 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29406301-hhnfj" event={"ID":"55460ba8-a9d1-4ef2-9bcf-e08d863698fa","Type":"ContainerDied","Data":"17757fc7390dc6deda9faf0a73c46419b174789f98c6fe90241f7d1f48e7047f"} Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.085232 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.118430 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-combined-ca-bundle\") pod \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.118593 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-fernet-keys\") pod \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.118764 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74nt6\" (UniqueName: \"kubernetes.io/projected/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-kube-api-access-74nt6\") pod \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.118856 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-config-data\") pod \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\" (UID: \"55460ba8-a9d1-4ef2-9bcf-e08d863698fa\") " Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.128893 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-kube-api-access-74nt6" (OuterVolumeSpecName: "kube-api-access-74nt6") pod "55460ba8-a9d1-4ef2-9bcf-e08d863698fa" (UID: "55460ba8-a9d1-4ef2-9bcf-e08d863698fa"). InnerVolumeSpecName "kube-api-access-74nt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.134880 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "55460ba8-a9d1-4ef2-9bcf-e08d863698fa" (UID: "55460ba8-a9d1-4ef2-9bcf-e08d863698fa"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.163088 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55460ba8-a9d1-4ef2-9bcf-e08d863698fa" (UID: "55460ba8-a9d1-4ef2-9bcf-e08d863698fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.178454 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-config-data" (OuterVolumeSpecName: "config-data") pod "55460ba8-a9d1-4ef2-9bcf-e08d863698fa" (UID: "55460ba8-a9d1-4ef2-9bcf-e08d863698fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.220950 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74nt6\" (UniqueName: \"kubernetes.io/projected/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-kube-api-access-74nt6\") on node \"crc\" DevicePath \"\"" Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.220982 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.220995 4931 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.221006 4931 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55460ba8-a9d1-4ef2-9bcf-e08d863698fa-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.580974 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29406301-hhnfj" event={"ID":"55460ba8-a9d1-4ef2-9bcf-e08d863698fa","Type":"ContainerDied","Data":"f2c65ce9c32b900b3000c2c4317d8357fbd48830bb550f2aa0a5ec62f7bde3ba"} Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.581332 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2c65ce9c32b900b3000c2c4317d8357fbd48830bb550f2aa0a5ec62f7bde3ba" Nov 29 01:01:05 crc kubenswrapper[4931]: I1129 01:01:05.581051 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29406301-hhnfj" Nov 29 01:02:53 crc kubenswrapper[4931]: I1129 01:02:53.735089 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 01:02:53 crc kubenswrapper[4931]: I1129 01:02:53.735937 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 01:03:23 crc kubenswrapper[4931]: I1129 01:03:23.735331 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 01:03:23 crc kubenswrapper[4931]: I1129 01:03:23.736176 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 01:03:53 crc kubenswrapper[4931]: I1129 01:03:53.735532 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 01:03:53 crc kubenswrapper[4931]: I1129 01:03:53.736156 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 01:03:53 crc kubenswrapper[4931]: I1129 01:03:53.736215 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 01:03:53 crc kubenswrapper[4931]: I1129 01:03:53.737090 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 01:03:53 crc kubenswrapper[4931]: I1129 01:03:53.737197 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" gracePeriod=600 Nov 29 01:03:53 crc kubenswrapper[4931]: E1129 01:03:53.869704 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:03:54 crc kubenswrapper[4931]: I1129 01:03:54.084774 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" exitCode=0 Nov 29 01:03:54 crc kubenswrapper[4931]: I1129 01:03:54.084897 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094"} Nov 29 01:03:54 crc kubenswrapper[4931]: I1129 01:03:54.085100 4931 scope.go:117] "RemoveContainer" containerID="089967d9016f33d2d57b978bd4d367473dd4604ca21d5078d10d82bb6e3d56fe" Nov 29 01:03:54 crc kubenswrapper[4931]: I1129 01:03:54.086109 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:03:54 crc kubenswrapper[4931]: E1129 01:03:54.086984 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:04:07 crc kubenswrapper[4931]: I1129 01:04:07.218352 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:04:07 crc kubenswrapper[4931]: E1129 01:04:07.219034 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:04:22 crc kubenswrapper[4931]: I1129 01:04:22.212902 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:04:22 crc kubenswrapper[4931]: E1129 01:04:22.216118 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:04:37 crc kubenswrapper[4931]: I1129 01:04:37.223764 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:04:37 crc kubenswrapper[4931]: E1129 01:04:37.224731 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:04:39 crc kubenswrapper[4931]: I1129 01:04:39.580419 4931 generic.go:334] "Generic (PLEG): container finished" podID="16c927ea-8729-4da8-8893-c23a53d85c5d" containerID="29f062c2ceeb95bc261f68a0d641884366dd6e33146c7a8b6627b1e7a7797c40" exitCode=0 Nov 29 01:04:39 crc kubenswrapper[4931]: I1129 01:04:39.580520 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"16c927ea-8729-4da8-8893-c23a53d85c5d","Type":"ContainerDied","Data":"29f062c2ceeb95bc261f68a0d641884366dd6e33146c7a8b6627b1e7a7797c40"} Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.117427 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.212796 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msmzv\" (UniqueName: \"kubernetes.io/projected/16c927ea-8729-4da8-8893-c23a53d85c5d-kube-api-access-msmzv\") pod \"16c927ea-8729-4da8-8893-c23a53d85c5d\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.213035 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-openstack-config-secret\") pod \"16c927ea-8729-4da8-8893-c23a53d85c5d\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.213098 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-ssh-key\") pod \"16c927ea-8729-4da8-8893-c23a53d85c5d\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.213222 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16c927ea-8729-4da8-8893-c23a53d85c5d-config-data\") pod \"16c927ea-8729-4da8-8893-c23a53d85c5d\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.213300 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"16c927ea-8729-4da8-8893-c23a53d85c5d\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.213365 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/16c927ea-8729-4da8-8893-c23a53d85c5d-test-operator-ephemeral-workdir\") pod \"16c927ea-8729-4da8-8893-c23a53d85c5d\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.213422 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/16c927ea-8729-4da8-8893-c23a53d85c5d-openstack-config\") pod \"16c927ea-8729-4da8-8893-c23a53d85c5d\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.213472 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-ca-certs\") pod \"16c927ea-8729-4da8-8893-c23a53d85c5d\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.213932 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/16c927ea-8729-4da8-8893-c23a53d85c5d-test-operator-ephemeral-temporary\") pod \"16c927ea-8729-4da8-8893-c23a53d85c5d\" (UID: \"16c927ea-8729-4da8-8893-c23a53d85c5d\") " Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.214362 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16c927ea-8729-4da8-8893-c23a53d85c5d-config-data" (OuterVolumeSpecName: "config-data") pod "16c927ea-8729-4da8-8893-c23a53d85c5d" (UID: "16c927ea-8729-4da8-8893-c23a53d85c5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.214446 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16c927ea-8729-4da8-8893-c23a53d85c5d-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "16c927ea-8729-4da8-8893-c23a53d85c5d" (UID: "16c927ea-8729-4da8-8893-c23a53d85c5d"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.215114 4931 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16c927ea-8729-4da8-8893-c23a53d85c5d-config-data\") on node \"crc\" DevicePath \"\"" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.215144 4931 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/16c927ea-8729-4da8-8893-c23a53d85c5d-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.216668 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16c927ea-8729-4da8-8893-c23a53d85c5d-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "16c927ea-8729-4da8-8893-c23a53d85c5d" (UID: "16c927ea-8729-4da8-8893-c23a53d85c5d"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.219126 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "16c927ea-8729-4da8-8893-c23a53d85c5d" (UID: "16c927ea-8729-4da8-8893-c23a53d85c5d"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.224102 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16c927ea-8729-4da8-8893-c23a53d85c5d-kube-api-access-msmzv" (OuterVolumeSpecName: "kube-api-access-msmzv") pod "16c927ea-8729-4da8-8893-c23a53d85c5d" (UID: "16c927ea-8729-4da8-8893-c23a53d85c5d"). InnerVolumeSpecName "kube-api-access-msmzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.244572 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "16c927ea-8729-4da8-8893-c23a53d85c5d" (UID: "16c927ea-8729-4da8-8893-c23a53d85c5d"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.245528 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "16c927ea-8729-4da8-8893-c23a53d85c5d" (UID: "16c927ea-8729-4da8-8893-c23a53d85c5d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.280500 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "16c927ea-8729-4da8-8893-c23a53d85c5d" (UID: "16c927ea-8729-4da8-8893-c23a53d85c5d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.282430 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16c927ea-8729-4da8-8893-c23a53d85c5d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "16c927ea-8729-4da8-8893-c23a53d85c5d" (UID: "16c927ea-8729-4da8-8893-c23a53d85c5d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.317847 4931 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/16c927ea-8729-4da8-8893-c23a53d85c5d-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.317898 4931 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/16c927ea-8729-4da8-8893-c23a53d85c5d-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.317921 4931 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.317958 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msmzv\" (UniqueName: \"kubernetes.io/projected/16c927ea-8729-4da8-8893-c23a53d85c5d-kube-api-access-msmzv\") on node \"crc\" DevicePath \"\"" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.317977 4931 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.318010 4931 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/16c927ea-8729-4da8-8893-c23a53d85c5d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.318046 4931 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.355675 4931 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.420314 4931 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.608052 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"16c927ea-8729-4da8-8893-c23a53d85c5d","Type":"ContainerDied","Data":"4e76f9057cfa66cfee9ceb419a2154295796ac12b981805e6dea9893373c2abb"} Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.608122 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e76f9057cfa66cfee9ceb419a2154295796ac12b981805e6dea9893373c2abb" Nov 29 01:04:41 crc kubenswrapper[4931]: I1129 01:04:41.608120 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.506067 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 29 01:04:49 crc kubenswrapper[4931]: E1129 01:04:49.507274 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c927ea-8729-4da8-8893-c23a53d85c5d" containerName="tempest-tests-tempest-tests-runner" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.507297 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c927ea-8729-4da8-8893-c23a53d85c5d" containerName="tempest-tests-tempest-tests-runner" Nov 29 01:04:49 crc kubenswrapper[4931]: E1129 01:04:49.507370 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55460ba8-a9d1-4ef2-9bcf-e08d863698fa" containerName="keystone-cron" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.507383 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="55460ba8-a9d1-4ef2-9bcf-e08d863698fa" containerName="keystone-cron" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.507724 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="16c927ea-8729-4da8-8893-c23a53d85c5d" containerName="tempest-tests-tempest-tests-runner" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.507760 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="55460ba8-a9d1-4ef2-9bcf-e08d863698fa" containerName="keystone-cron" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.508871 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.522912 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.541133 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-gnq5t" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.624548 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6jlm\" (UniqueName: \"kubernetes.io/projected/dbcaae8b-9df7-4dda-a4ea-3a237c20627f-kube-api-access-p6jlm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dbcaae8b-9df7-4dda-a4ea-3a237c20627f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.624616 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dbcaae8b-9df7-4dda-a4ea-3a237c20627f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.726604 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6jlm\" (UniqueName: \"kubernetes.io/projected/dbcaae8b-9df7-4dda-a4ea-3a237c20627f-kube-api-access-p6jlm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dbcaae8b-9df7-4dda-a4ea-3a237c20627f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.727015 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dbcaae8b-9df7-4dda-a4ea-3a237c20627f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.727671 4931 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dbcaae8b-9df7-4dda-a4ea-3a237c20627f\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.749374 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6jlm\" (UniqueName: \"kubernetes.io/projected/dbcaae8b-9df7-4dda-a4ea-3a237c20627f-kube-api-access-p6jlm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dbcaae8b-9df7-4dda-a4ea-3a237c20627f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.764903 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dbcaae8b-9df7-4dda-a4ea-3a237c20627f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 01:04:49 crc kubenswrapper[4931]: I1129 01:04:49.870078 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 29 01:04:50 crc kubenswrapper[4931]: I1129 01:04:50.371626 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 29 01:04:50 crc kubenswrapper[4931]: I1129 01:04:50.379516 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 01:04:50 crc kubenswrapper[4931]: I1129 01:04:50.728302 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"dbcaae8b-9df7-4dda-a4ea-3a237c20627f","Type":"ContainerStarted","Data":"50bcadc5de5b5985f400ed561de2ac8d118d99397e9c71f318ab2bb159fc4c18"} Nov 29 01:04:51 crc kubenswrapper[4931]: I1129 01:04:51.743017 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"dbcaae8b-9df7-4dda-a4ea-3a237c20627f","Type":"ContainerStarted","Data":"2ea8f6136ac3e44534ea68d66be0b936db004d33017de523a2d02805ee4c3a42"} Nov 29 01:04:51 crc kubenswrapper[4931]: I1129 01:04:51.772893 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.750449142 podStartE2EDuration="2.772871776s" podCreationTimestamp="2025-11-29 01:04:49 +0000 UTC" firstStartedPulling="2025-11-29 01:04:50.379053449 +0000 UTC m=+3903.540946721" lastFinishedPulling="2025-11-29 01:04:51.401476083 +0000 UTC m=+3904.563369355" observedRunningTime="2025-11-29 01:04:51.761243847 +0000 UTC m=+3904.923137109" watchObservedRunningTime="2025-11-29 01:04:51.772871776 +0000 UTC m=+3904.934765018" Nov 29 01:04:52 crc kubenswrapper[4931]: I1129 01:04:52.214576 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:04:52 crc kubenswrapper[4931]: E1129 01:04:52.217783 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:05:07 crc kubenswrapper[4931]: I1129 01:05:07.226252 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:05:07 crc kubenswrapper[4931]: E1129 01:05:07.227373 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.431462 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wn64l/must-gather-lht9f"] Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.433309 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/must-gather-lht9f" Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.440673 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wn64l/must-gather-lht9f"] Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.441437 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-wn64l"/"default-dockercfg-rc79t" Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.441549 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wn64l"/"kube-root-ca.crt" Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.441611 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wn64l"/"openshift-service-ca.crt" Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.479138 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/12d5edce-1a2f-4cce-89a4-99ed66c0cc15-must-gather-output\") pod \"must-gather-lht9f\" (UID: \"12d5edce-1a2f-4cce-89a4-99ed66c0cc15\") " pod="openshift-must-gather-wn64l/must-gather-lht9f" Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.479267 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqwk6\" (UniqueName: \"kubernetes.io/projected/12d5edce-1a2f-4cce-89a4-99ed66c0cc15-kube-api-access-qqwk6\") pod \"must-gather-lht9f\" (UID: \"12d5edce-1a2f-4cce-89a4-99ed66c0cc15\") " pod="openshift-must-gather-wn64l/must-gather-lht9f" Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.580760 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/12d5edce-1a2f-4cce-89a4-99ed66c0cc15-must-gather-output\") pod \"must-gather-lht9f\" (UID: \"12d5edce-1a2f-4cce-89a4-99ed66c0cc15\") " pod="openshift-must-gather-wn64l/must-gather-lht9f" Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.580882 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqwk6\" (UniqueName: \"kubernetes.io/projected/12d5edce-1a2f-4cce-89a4-99ed66c0cc15-kube-api-access-qqwk6\") pod \"must-gather-lht9f\" (UID: \"12d5edce-1a2f-4cce-89a4-99ed66c0cc15\") " pod="openshift-must-gather-wn64l/must-gather-lht9f" Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.581285 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/12d5edce-1a2f-4cce-89a4-99ed66c0cc15-must-gather-output\") pod \"must-gather-lht9f\" (UID: \"12d5edce-1a2f-4cce-89a4-99ed66c0cc15\") " pod="openshift-must-gather-wn64l/must-gather-lht9f" Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.600450 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqwk6\" (UniqueName: \"kubernetes.io/projected/12d5edce-1a2f-4cce-89a4-99ed66c0cc15-kube-api-access-qqwk6\") pod \"must-gather-lht9f\" (UID: \"12d5edce-1a2f-4cce-89a4-99ed66c0cc15\") " pod="openshift-must-gather-wn64l/must-gather-lht9f" Nov 29 01:05:14 crc kubenswrapper[4931]: I1129 01:05:14.760104 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/must-gather-lht9f" Nov 29 01:05:15 crc kubenswrapper[4931]: I1129 01:05:15.295532 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wn64l/must-gather-lht9f"] Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.026798 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wn64l/must-gather-lht9f" event={"ID":"12d5edce-1a2f-4cce-89a4-99ed66c0cc15","Type":"ContainerStarted","Data":"45b4afb589a71172be3f5f134d6ba8669b98130d6ac57e1fc19431b17e510bcb"} Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.081904 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4lqsq"] Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.084743 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.093402 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lqsq"] Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.218783 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3244f88-89cd-43be-95a9-38fd5d685e5b-catalog-content\") pod \"redhat-marketplace-4lqsq\" (UID: \"d3244f88-89cd-43be-95a9-38fd5d685e5b\") " pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.218878 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9fdh\" (UniqueName: \"kubernetes.io/projected/d3244f88-89cd-43be-95a9-38fd5d685e5b-kube-api-access-t9fdh\") pod \"redhat-marketplace-4lqsq\" (UID: \"d3244f88-89cd-43be-95a9-38fd5d685e5b\") " pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.218902 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3244f88-89cd-43be-95a9-38fd5d685e5b-utilities\") pod \"redhat-marketplace-4lqsq\" (UID: \"d3244f88-89cd-43be-95a9-38fd5d685e5b\") " pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.320372 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3244f88-89cd-43be-95a9-38fd5d685e5b-catalog-content\") pod \"redhat-marketplace-4lqsq\" (UID: \"d3244f88-89cd-43be-95a9-38fd5d685e5b\") " pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.320453 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9fdh\" (UniqueName: \"kubernetes.io/projected/d3244f88-89cd-43be-95a9-38fd5d685e5b-kube-api-access-t9fdh\") pod \"redhat-marketplace-4lqsq\" (UID: \"d3244f88-89cd-43be-95a9-38fd5d685e5b\") " pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.320476 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3244f88-89cd-43be-95a9-38fd5d685e5b-utilities\") pod \"redhat-marketplace-4lqsq\" (UID: \"d3244f88-89cd-43be-95a9-38fd5d685e5b\") " pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.321103 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3244f88-89cd-43be-95a9-38fd5d685e5b-utilities\") pod \"redhat-marketplace-4lqsq\" (UID: \"d3244f88-89cd-43be-95a9-38fd5d685e5b\") " pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.321340 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3244f88-89cd-43be-95a9-38fd5d685e5b-catalog-content\") pod \"redhat-marketplace-4lqsq\" (UID: \"d3244f88-89cd-43be-95a9-38fd5d685e5b\") " pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.347037 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9fdh\" (UniqueName: \"kubernetes.io/projected/d3244f88-89cd-43be-95a9-38fd5d685e5b-kube-api-access-t9fdh\") pod \"redhat-marketplace-4lqsq\" (UID: \"d3244f88-89cd-43be-95a9-38fd5d685e5b\") " pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.412099 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:16 crc kubenswrapper[4931]: I1129 01:05:16.933673 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lqsq"] Nov 29 01:05:17 crc kubenswrapper[4931]: I1129 01:05:17.040051 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lqsq" event={"ID":"d3244f88-89cd-43be-95a9-38fd5d685e5b","Type":"ContainerStarted","Data":"d89e1146693e3c4696e07ebdc6ad13483817437cf72ca74af8565c6d2a58147a"} Nov 29 01:05:18 crc kubenswrapper[4931]: I1129 01:05:18.051927 4931 generic.go:334] "Generic (PLEG): container finished" podID="d3244f88-89cd-43be-95a9-38fd5d685e5b" containerID="ffc52ece88cbd45828a306feb4be15c95e4a88374b8d15585754f0cf23df7d91" exitCode=0 Nov 29 01:05:18 crc kubenswrapper[4931]: I1129 01:05:18.052409 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lqsq" event={"ID":"d3244f88-89cd-43be-95a9-38fd5d685e5b","Type":"ContainerDied","Data":"ffc52ece88cbd45828a306feb4be15c95e4a88374b8d15585754f0cf23df7d91"} Nov 29 01:05:18 crc kubenswrapper[4931]: I1129 01:05:18.214062 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:05:18 crc kubenswrapper[4931]: E1129 01:05:18.214325 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:05:20 crc kubenswrapper[4931]: I1129 01:05:20.083548 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wn64l/must-gather-lht9f" event={"ID":"12d5edce-1a2f-4cce-89a4-99ed66c0cc15","Type":"ContainerStarted","Data":"6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336"} Nov 29 01:05:21 crc kubenswrapper[4931]: I1129 01:05:21.107309 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wn64l/must-gather-lht9f" event={"ID":"12d5edce-1a2f-4cce-89a4-99ed66c0cc15","Type":"ContainerStarted","Data":"e0021ea63156d65ae84861c442c97a45bc1771a869423dd4cb8b53ca9edd0ced"} Nov 29 01:05:21 crc kubenswrapper[4931]: I1129 01:05:21.111062 4931 generic.go:334] "Generic (PLEG): container finished" podID="d3244f88-89cd-43be-95a9-38fd5d685e5b" containerID="a0ae6b8a93b24e24361082cbf7f887cd5ce1e65f8e914a122192534e95ab30a6" exitCode=0 Nov 29 01:05:21 crc kubenswrapper[4931]: I1129 01:05:21.111102 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lqsq" event={"ID":"d3244f88-89cd-43be-95a9-38fd5d685e5b","Type":"ContainerDied","Data":"a0ae6b8a93b24e24361082cbf7f887cd5ce1e65f8e914a122192534e95ab30a6"} Nov 29 01:05:21 crc kubenswrapper[4931]: I1129 01:05:21.137307 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wn64l/must-gather-lht9f" podStartSLOduration=2.7705267239999998 podStartE2EDuration="7.137287302s" podCreationTimestamp="2025-11-29 01:05:14 +0000 UTC" firstStartedPulling="2025-11-29 01:05:15.304688551 +0000 UTC m=+3928.466581783" lastFinishedPulling="2025-11-29 01:05:19.671449129 +0000 UTC m=+3932.833342361" observedRunningTime="2025-11-29 01:05:21.126302292 +0000 UTC m=+3934.288195524" watchObservedRunningTime="2025-11-29 01:05:21.137287302 +0000 UTC m=+3934.299180534" Nov 29 01:05:22 crc kubenswrapper[4931]: I1129 01:05:22.126935 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lqsq" event={"ID":"d3244f88-89cd-43be-95a9-38fd5d685e5b","Type":"ContainerStarted","Data":"e53aebfd8e342e96f6fe27e69b1119915ce49f5dea78cac8b2df4a0afc26a1ea"} Nov 29 01:05:22 crc kubenswrapper[4931]: I1129 01:05:22.152440 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4lqsq" podStartSLOduration=4.10679723 podStartE2EDuration="6.152426032s" podCreationTimestamp="2025-11-29 01:05:16 +0000 UTC" firstStartedPulling="2025-11-29 01:05:19.604716865 +0000 UTC m=+3932.766610097" lastFinishedPulling="2025-11-29 01:05:21.650345657 +0000 UTC m=+3934.812238899" observedRunningTime="2025-11-29 01:05:22.143044437 +0000 UTC m=+3935.304937669" watchObservedRunningTime="2025-11-29 01:05:22.152426032 +0000 UTC m=+3935.314319264" Nov 29 01:05:23 crc kubenswrapper[4931]: I1129 01:05:23.537563 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wn64l/crc-debug-f97dw"] Nov 29 01:05:23 crc kubenswrapper[4931]: I1129 01:05:23.538900 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/crc-debug-f97dw" Nov 29 01:05:23 crc kubenswrapper[4931]: I1129 01:05:23.663761 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a634d50-5cc2-47b8-a70a-86635250b50a-host\") pod \"crc-debug-f97dw\" (UID: \"9a634d50-5cc2-47b8-a70a-86635250b50a\") " pod="openshift-must-gather-wn64l/crc-debug-f97dw" Nov 29 01:05:23 crc kubenswrapper[4931]: I1129 01:05:23.664613 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfhk7\" (UniqueName: \"kubernetes.io/projected/9a634d50-5cc2-47b8-a70a-86635250b50a-kube-api-access-mfhk7\") pod \"crc-debug-f97dw\" (UID: \"9a634d50-5cc2-47b8-a70a-86635250b50a\") " pod="openshift-must-gather-wn64l/crc-debug-f97dw" Nov 29 01:05:23 crc kubenswrapper[4931]: I1129 01:05:23.766909 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfhk7\" (UniqueName: \"kubernetes.io/projected/9a634d50-5cc2-47b8-a70a-86635250b50a-kube-api-access-mfhk7\") pod \"crc-debug-f97dw\" (UID: \"9a634d50-5cc2-47b8-a70a-86635250b50a\") " pod="openshift-must-gather-wn64l/crc-debug-f97dw" Nov 29 01:05:23 crc kubenswrapper[4931]: I1129 01:05:23.767169 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a634d50-5cc2-47b8-a70a-86635250b50a-host\") pod \"crc-debug-f97dw\" (UID: \"9a634d50-5cc2-47b8-a70a-86635250b50a\") " pod="openshift-must-gather-wn64l/crc-debug-f97dw" Nov 29 01:05:23 crc kubenswrapper[4931]: I1129 01:05:23.767285 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a634d50-5cc2-47b8-a70a-86635250b50a-host\") pod \"crc-debug-f97dw\" (UID: \"9a634d50-5cc2-47b8-a70a-86635250b50a\") " pod="openshift-must-gather-wn64l/crc-debug-f97dw" Nov 29 01:05:23 crc kubenswrapper[4931]: I1129 01:05:23.784986 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfhk7\" (UniqueName: \"kubernetes.io/projected/9a634d50-5cc2-47b8-a70a-86635250b50a-kube-api-access-mfhk7\") pod \"crc-debug-f97dw\" (UID: \"9a634d50-5cc2-47b8-a70a-86635250b50a\") " pod="openshift-must-gather-wn64l/crc-debug-f97dw" Nov 29 01:05:23 crc kubenswrapper[4931]: I1129 01:05:23.863071 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/crc-debug-f97dw" Nov 29 01:05:23 crc kubenswrapper[4931]: W1129 01:05:23.888689 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a634d50_5cc2_47b8_a70a_86635250b50a.slice/crio-aba79cf64cddc02bd4630d762b277089935ea4df4fc7d67d157b06ea54f9c5b5 WatchSource:0}: Error finding container aba79cf64cddc02bd4630d762b277089935ea4df4fc7d67d157b06ea54f9c5b5: Status 404 returned error can't find the container with id aba79cf64cddc02bd4630d762b277089935ea4df4fc7d67d157b06ea54f9c5b5 Nov 29 01:05:24 crc kubenswrapper[4931]: I1129 01:05:24.144130 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wn64l/crc-debug-f97dw" event={"ID":"9a634d50-5cc2-47b8-a70a-86635250b50a","Type":"ContainerStarted","Data":"aba79cf64cddc02bd4630d762b277089935ea4df4fc7d67d157b06ea54f9c5b5"} Nov 29 01:05:26 crc kubenswrapper[4931]: I1129 01:05:26.412580 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:26 crc kubenswrapper[4931]: I1129 01:05:26.413242 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:26 crc kubenswrapper[4931]: I1129 01:05:26.467585 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:27 crc kubenswrapper[4931]: I1129 01:05:27.222512 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:27 crc kubenswrapper[4931]: I1129 01:05:27.279527 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lqsq"] Nov 29 01:05:29 crc kubenswrapper[4931]: I1129 01:05:29.186923 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4lqsq" podUID="d3244f88-89cd-43be-95a9-38fd5d685e5b" containerName="registry-server" containerID="cri-o://e53aebfd8e342e96f6fe27e69b1119915ce49f5dea78cac8b2df4a0afc26a1ea" gracePeriod=2 Nov 29 01:05:29 crc kubenswrapper[4931]: I1129 01:05:29.213336 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:05:29 crc kubenswrapper[4931]: E1129 01:05:29.213535 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:05:30 crc kubenswrapper[4931]: I1129 01:05:30.196900 4931 generic.go:334] "Generic (PLEG): container finished" podID="d3244f88-89cd-43be-95a9-38fd5d685e5b" containerID="e53aebfd8e342e96f6fe27e69b1119915ce49f5dea78cac8b2df4a0afc26a1ea" exitCode=0 Nov 29 01:05:30 crc kubenswrapper[4931]: I1129 01:05:30.196977 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lqsq" event={"ID":"d3244f88-89cd-43be-95a9-38fd5d685e5b","Type":"ContainerDied","Data":"e53aebfd8e342e96f6fe27e69b1119915ce49f5dea78cac8b2df4a0afc26a1ea"} Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.244107 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.250948 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wn64l/crc-debug-f97dw" event={"ID":"9a634d50-5cc2-47b8-a70a-86635250b50a","Type":"ContainerStarted","Data":"0472fcbea5dcb603175d817c890e59959b3eb3d1b2f1c54321630e9e5e0800dd"} Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.253150 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lqsq" event={"ID":"d3244f88-89cd-43be-95a9-38fd5d685e5b","Type":"ContainerDied","Data":"d89e1146693e3c4696e07ebdc6ad13483817437cf72ca74af8565c6d2a58147a"} Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.253191 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lqsq" Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.253195 4931 scope.go:117] "RemoveContainer" containerID="e53aebfd8e342e96f6fe27e69b1119915ce49f5dea78cac8b2df4a0afc26a1ea" Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.280517 4931 scope.go:117] "RemoveContainer" containerID="a0ae6b8a93b24e24361082cbf7f887cd5ce1e65f8e914a122192534e95ab30a6" Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.296999 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wn64l/crc-debug-f97dw" podStartSLOduration=1.226730845 podStartE2EDuration="12.296980439s" podCreationTimestamp="2025-11-29 01:05:23 +0000 UTC" firstStartedPulling="2025-11-29 01:05:23.890706591 +0000 UTC m=+3937.052599823" lastFinishedPulling="2025-11-29 01:05:34.960956185 +0000 UTC m=+3948.122849417" observedRunningTime="2025-11-29 01:05:35.286590005 +0000 UTC m=+3948.448483237" watchObservedRunningTime="2025-11-29 01:05:35.296980439 +0000 UTC m=+3948.458873671" Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.351191 4931 scope.go:117] "RemoveContainer" containerID="ffc52ece88cbd45828a306feb4be15c95e4a88374b8d15585754f0cf23df7d91" Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.377767 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3244f88-89cd-43be-95a9-38fd5d685e5b-utilities\") pod \"d3244f88-89cd-43be-95a9-38fd5d685e5b\" (UID: \"d3244f88-89cd-43be-95a9-38fd5d685e5b\") " Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.377900 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3244f88-89cd-43be-95a9-38fd5d685e5b-catalog-content\") pod \"d3244f88-89cd-43be-95a9-38fd5d685e5b\" (UID: \"d3244f88-89cd-43be-95a9-38fd5d685e5b\") " Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.378098 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9fdh\" (UniqueName: \"kubernetes.io/projected/d3244f88-89cd-43be-95a9-38fd5d685e5b-kube-api-access-t9fdh\") pod \"d3244f88-89cd-43be-95a9-38fd5d685e5b\" (UID: \"d3244f88-89cd-43be-95a9-38fd5d685e5b\") " Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.378769 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3244f88-89cd-43be-95a9-38fd5d685e5b-utilities" (OuterVolumeSpecName: "utilities") pod "d3244f88-89cd-43be-95a9-38fd5d685e5b" (UID: "d3244f88-89cd-43be-95a9-38fd5d685e5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.383681 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3244f88-89cd-43be-95a9-38fd5d685e5b-kube-api-access-t9fdh" (OuterVolumeSpecName: "kube-api-access-t9fdh") pod "d3244f88-89cd-43be-95a9-38fd5d685e5b" (UID: "d3244f88-89cd-43be-95a9-38fd5d685e5b"). InnerVolumeSpecName "kube-api-access-t9fdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.396480 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3244f88-89cd-43be-95a9-38fd5d685e5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3244f88-89cd-43be-95a9-38fd5d685e5b" (UID: "d3244f88-89cd-43be-95a9-38fd5d685e5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.480173 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3244f88-89cd-43be-95a9-38fd5d685e5b-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.480417 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3244f88-89cd-43be-95a9-38fd5d685e5b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.480429 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9fdh\" (UniqueName: \"kubernetes.io/projected/d3244f88-89cd-43be-95a9-38fd5d685e5b-kube-api-access-t9fdh\") on node \"crc\" DevicePath \"\"" Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.620194 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lqsq"] Nov 29 01:05:35 crc kubenswrapper[4931]: I1129 01:05:35.654180 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lqsq"] Nov 29 01:05:37 crc kubenswrapper[4931]: I1129 01:05:37.224116 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3244f88-89cd-43be-95a9-38fd5d685e5b" path="/var/lib/kubelet/pods/d3244f88-89cd-43be-95a9-38fd5d685e5b/volumes" Nov 29 01:05:40 crc kubenswrapper[4931]: I1129 01:05:40.212555 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:05:40 crc kubenswrapper[4931]: E1129 01:05:40.213289 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:05:53 crc kubenswrapper[4931]: I1129 01:05:53.212778 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:05:53 crc kubenswrapper[4931]: E1129 01:05:53.213581 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:06:08 crc kubenswrapper[4931]: I1129 01:06:08.212723 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:06:08 crc kubenswrapper[4931]: E1129 01:06:08.213424 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:06:11 crc kubenswrapper[4931]: I1129 01:06:11.615916 4931 generic.go:334] "Generic (PLEG): container finished" podID="9a634d50-5cc2-47b8-a70a-86635250b50a" containerID="0472fcbea5dcb603175d817c890e59959b3eb3d1b2f1c54321630e9e5e0800dd" exitCode=0 Nov 29 01:06:11 crc kubenswrapper[4931]: I1129 01:06:11.615956 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wn64l/crc-debug-f97dw" event={"ID":"9a634d50-5cc2-47b8-a70a-86635250b50a","Type":"ContainerDied","Data":"0472fcbea5dcb603175d817c890e59959b3eb3d1b2f1c54321630e9e5e0800dd"} Nov 29 01:06:13 crc kubenswrapper[4931]: I1129 01:06:13.496668 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/crc-debug-f97dw" Nov 29 01:06:13 crc kubenswrapper[4931]: I1129 01:06:13.536319 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wn64l/crc-debug-f97dw"] Nov 29 01:06:13 crc kubenswrapper[4931]: I1129 01:06:13.545771 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wn64l/crc-debug-f97dw"] Nov 29 01:06:13 crc kubenswrapper[4931]: I1129 01:06:13.637746 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfhk7\" (UniqueName: \"kubernetes.io/projected/9a634d50-5cc2-47b8-a70a-86635250b50a-kube-api-access-mfhk7\") pod \"9a634d50-5cc2-47b8-a70a-86635250b50a\" (UID: \"9a634d50-5cc2-47b8-a70a-86635250b50a\") " Nov 29 01:06:13 crc kubenswrapper[4931]: I1129 01:06:13.638001 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a634d50-5cc2-47b8-a70a-86635250b50a-host\") pod \"9a634d50-5cc2-47b8-a70a-86635250b50a\" (UID: \"9a634d50-5cc2-47b8-a70a-86635250b50a\") " Nov 29 01:06:13 crc kubenswrapper[4931]: I1129 01:06:13.638135 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a634d50-5cc2-47b8-a70a-86635250b50a-host" (OuterVolumeSpecName: "host") pod "9a634d50-5cc2-47b8-a70a-86635250b50a" (UID: "9a634d50-5cc2-47b8-a70a-86635250b50a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 01:06:13 crc kubenswrapper[4931]: I1129 01:06:13.638743 4931 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a634d50-5cc2-47b8-a70a-86635250b50a-host\") on node \"crc\" DevicePath \"\"" Nov 29 01:06:13 crc kubenswrapper[4931]: I1129 01:06:13.645402 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a634d50-5cc2-47b8-a70a-86635250b50a-kube-api-access-mfhk7" (OuterVolumeSpecName: "kube-api-access-mfhk7") pod "9a634d50-5cc2-47b8-a70a-86635250b50a" (UID: "9a634d50-5cc2-47b8-a70a-86635250b50a"). InnerVolumeSpecName "kube-api-access-mfhk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:06:13 crc kubenswrapper[4931]: I1129 01:06:13.650768 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aba79cf64cddc02bd4630d762b277089935ea4df4fc7d67d157b06ea54f9c5b5" Nov 29 01:06:13 crc kubenswrapper[4931]: I1129 01:06:13.650872 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/crc-debug-f97dw" Nov 29 01:06:13 crc kubenswrapper[4931]: I1129 01:06:13.739575 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfhk7\" (UniqueName: \"kubernetes.io/projected/9a634d50-5cc2-47b8-a70a-86635250b50a-kube-api-access-mfhk7\") on node \"crc\" DevicePath \"\"" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.806376 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wn64l/crc-debug-tmjkz"] Nov 29 01:06:14 crc kubenswrapper[4931]: E1129 01:06:14.807940 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3244f88-89cd-43be-95a9-38fd5d685e5b" containerName="extract-content" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.807963 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3244f88-89cd-43be-95a9-38fd5d685e5b" containerName="extract-content" Nov 29 01:06:14 crc kubenswrapper[4931]: E1129 01:06:14.807999 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a634d50-5cc2-47b8-a70a-86635250b50a" containerName="container-00" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.808010 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a634d50-5cc2-47b8-a70a-86635250b50a" containerName="container-00" Nov 29 01:06:14 crc kubenswrapper[4931]: E1129 01:06:14.808024 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3244f88-89cd-43be-95a9-38fd5d685e5b" containerName="registry-server" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.808034 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3244f88-89cd-43be-95a9-38fd5d685e5b" containerName="registry-server" Nov 29 01:06:14 crc kubenswrapper[4931]: E1129 01:06:14.808063 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3244f88-89cd-43be-95a9-38fd5d685e5b" containerName="extract-utilities" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.808074 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3244f88-89cd-43be-95a9-38fd5d685e5b" containerName="extract-utilities" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.808366 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a634d50-5cc2-47b8-a70a-86635250b50a" containerName="container-00" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.808387 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3244f88-89cd-43be-95a9-38fd5d685e5b" containerName="registry-server" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.809248 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/crc-debug-tmjkz" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.866085 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs2dk\" (UniqueName: \"kubernetes.io/projected/a1d75ee8-9f23-4c05-9811-79bd1ef1a928-kube-api-access-cs2dk\") pod \"crc-debug-tmjkz\" (UID: \"a1d75ee8-9f23-4c05-9811-79bd1ef1a928\") " pod="openshift-must-gather-wn64l/crc-debug-tmjkz" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.866158 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1d75ee8-9f23-4c05-9811-79bd1ef1a928-host\") pod \"crc-debug-tmjkz\" (UID: \"a1d75ee8-9f23-4c05-9811-79bd1ef1a928\") " pod="openshift-must-gather-wn64l/crc-debug-tmjkz" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.969112 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs2dk\" (UniqueName: \"kubernetes.io/projected/a1d75ee8-9f23-4c05-9811-79bd1ef1a928-kube-api-access-cs2dk\") pod \"crc-debug-tmjkz\" (UID: \"a1d75ee8-9f23-4c05-9811-79bd1ef1a928\") " pod="openshift-must-gather-wn64l/crc-debug-tmjkz" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.969180 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1d75ee8-9f23-4c05-9811-79bd1ef1a928-host\") pod \"crc-debug-tmjkz\" (UID: \"a1d75ee8-9f23-4c05-9811-79bd1ef1a928\") " pod="openshift-must-gather-wn64l/crc-debug-tmjkz" Nov 29 01:06:14 crc kubenswrapper[4931]: I1129 01:06:14.969401 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1d75ee8-9f23-4c05-9811-79bd1ef1a928-host\") pod \"crc-debug-tmjkz\" (UID: \"a1d75ee8-9f23-4c05-9811-79bd1ef1a928\") " pod="openshift-must-gather-wn64l/crc-debug-tmjkz" Nov 29 01:06:15 crc kubenswrapper[4931]: I1129 01:06:15.000218 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs2dk\" (UniqueName: \"kubernetes.io/projected/a1d75ee8-9f23-4c05-9811-79bd1ef1a928-kube-api-access-cs2dk\") pod \"crc-debug-tmjkz\" (UID: \"a1d75ee8-9f23-4c05-9811-79bd1ef1a928\") " pod="openshift-must-gather-wn64l/crc-debug-tmjkz" Nov 29 01:06:15 crc kubenswrapper[4931]: I1129 01:06:15.131577 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/crc-debug-tmjkz" Nov 29 01:06:15 crc kubenswrapper[4931]: I1129 01:06:15.238587 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a634d50-5cc2-47b8-a70a-86635250b50a" path="/var/lib/kubelet/pods/9a634d50-5cc2-47b8-a70a-86635250b50a/volumes" Nov 29 01:06:15 crc kubenswrapper[4931]: I1129 01:06:15.677516 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wn64l/crc-debug-tmjkz" event={"ID":"a1d75ee8-9f23-4c05-9811-79bd1ef1a928","Type":"ContainerStarted","Data":"8818b1eccdb6eeb70d0918f7c425a42970a1da8fa6f089e2b38e83104a93e428"} Nov 29 01:06:16 crc kubenswrapper[4931]: I1129 01:06:16.692152 4931 generic.go:334] "Generic (PLEG): container finished" podID="a1d75ee8-9f23-4c05-9811-79bd1ef1a928" containerID="8442c11661395f41812caac6245b65f96cfb41e410279df40f9e124140fe848a" exitCode=0 Nov 29 01:06:16 crc kubenswrapper[4931]: I1129 01:06:16.692215 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wn64l/crc-debug-tmjkz" event={"ID":"a1d75ee8-9f23-4c05-9811-79bd1ef1a928","Type":"ContainerDied","Data":"8442c11661395f41812caac6245b65f96cfb41e410279df40f9e124140fe848a"} Nov 29 01:06:17 crc kubenswrapper[4931]: I1129 01:06:17.340684 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wn64l/crc-debug-tmjkz"] Nov 29 01:06:17 crc kubenswrapper[4931]: I1129 01:06:17.347932 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wn64l/crc-debug-tmjkz"] Nov 29 01:06:17 crc kubenswrapper[4931]: I1129 01:06:17.834704 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/crc-debug-tmjkz" Nov 29 01:06:17 crc kubenswrapper[4931]: I1129 01:06:17.926969 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs2dk\" (UniqueName: \"kubernetes.io/projected/a1d75ee8-9f23-4c05-9811-79bd1ef1a928-kube-api-access-cs2dk\") pod \"a1d75ee8-9f23-4c05-9811-79bd1ef1a928\" (UID: \"a1d75ee8-9f23-4c05-9811-79bd1ef1a928\") " Nov 29 01:06:17 crc kubenswrapper[4931]: I1129 01:06:17.927086 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1d75ee8-9f23-4c05-9811-79bd1ef1a928-host\") pod \"a1d75ee8-9f23-4c05-9811-79bd1ef1a928\" (UID: \"a1d75ee8-9f23-4c05-9811-79bd1ef1a928\") " Nov 29 01:06:17 crc kubenswrapper[4931]: I1129 01:06:17.927208 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1d75ee8-9f23-4c05-9811-79bd1ef1a928-host" (OuterVolumeSpecName: "host") pod "a1d75ee8-9f23-4c05-9811-79bd1ef1a928" (UID: "a1d75ee8-9f23-4c05-9811-79bd1ef1a928"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 01:06:17 crc kubenswrapper[4931]: I1129 01:06:17.927948 4931 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a1d75ee8-9f23-4c05-9811-79bd1ef1a928-host\") on node \"crc\" DevicePath \"\"" Nov 29 01:06:17 crc kubenswrapper[4931]: I1129 01:06:17.935446 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1d75ee8-9f23-4c05-9811-79bd1ef1a928-kube-api-access-cs2dk" (OuterVolumeSpecName: "kube-api-access-cs2dk") pod "a1d75ee8-9f23-4c05-9811-79bd1ef1a928" (UID: "a1d75ee8-9f23-4c05-9811-79bd1ef1a928"). InnerVolumeSpecName "kube-api-access-cs2dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.030798 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs2dk\" (UniqueName: \"kubernetes.io/projected/a1d75ee8-9f23-4c05-9811-79bd1ef1a928-kube-api-access-cs2dk\") on node \"crc\" DevicePath \"\"" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.548567 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wn64l/crc-debug-hf6p2"] Nov 29 01:06:18 crc kubenswrapper[4931]: E1129 01:06:18.550865 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1d75ee8-9f23-4c05-9811-79bd1ef1a928" containerName="container-00" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.550919 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1d75ee8-9f23-4c05-9811-79bd1ef1a928" containerName="container-00" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.552095 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1d75ee8-9f23-4c05-9811-79bd1ef1a928" containerName="container-00" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.553625 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/crc-debug-hf6p2" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.642218 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/93130af3-0a61-4f95-8eba-36321928d128-host\") pod \"crc-debug-hf6p2\" (UID: \"93130af3-0a61-4f95-8eba-36321928d128\") " pod="openshift-must-gather-wn64l/crc-debug-hf6p2" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.642573 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tb5k\" (UniqueName: \"kubernetes.io/projected/93130af3-0a61-4f95-8eba-36321928d128-kube-api-access-8tb5k\") pod \"crc-debug-hf6p2\" (UID: \"93130af3-0a61-4f95-8eba-36321928d128\") " pod="openshift-must-gather-wn64l/crc-debug-hf6p2" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.717588 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8818b1eccdb6eeb70d0918f7c425a42970a1da8fa6f089e2b38e83104a93e428" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.717624 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/crc-debug-tmjkz" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.744481 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/93130af3-0a61-4f95-8eba-36321928d128-host\") pod \"crc-debug-hf6p2\" (UID: \"93130af3-0a61-4f95-8eba-36321928d128\") " pod="openshift-must-gather-wn64l/crc-debug-hf6p2" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.744645 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/93130af3-0a61-4f95-8eba-36321928d128-host\") pod \"crc-debug-hf6p2\" (UID: \"93130af3-0a61-4f95-8eba-36321928d128\") " pod="openshift-must-gather-wn64l/crc-debug-hf6p2" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.744875 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tb5k\" (UniqueName: \"kubernetes.io/projected/93130af3-0a61-4f95-8eba-36321928d128-kube-api-access-8tb5k\") pod \"crc-debug-hf6p2\" (UID: \"93130af3-0a61-4f95-8eba-36321928d128\") " pod="openshift-must-gather-wn64l/crc-debug-hf6p2" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.767578 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tb5k\" (UniqueName: \"kubernetes.io/projected/93130af3-0a61-4f95-8eba-36321928d128-kube-api-access-8tb5k\") pod \"crc-debug-hf6p2\" (UID: \"93130af3-0a61-4f95-8eba-36321928d128\") " pod="openshift-must-gather-wn64l/crc-debug-hf6p2" Nov 29 01:06:18 crc kubenswrapper[4931]: I1129 01:06:18.881435 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/crc-debug-hf6p2" Nov 29 01:06:19 crc kubenswrapper[4931]: I1129 01:06:19.231619 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1d75ee8-9f23-4c05-9811-79bd1ef1a928" path="/var/lib/kubelet/pods/a1d75ee8-9f23-4c05-9811-79bd1ef1a928/volumes" Nov 29 01:06:19 crc kubenswrapper[4931]: I1129 01:06:19.735782 4931 generic.go:334] "Generic (PLEG): container finished" podID="93130af3-0a61-4f95-8eba-36321928d128" containerID="d2d540418c4d2e80a6bd5cf4b37fd296c7ac883630b18a876c663e2b4e6517c4" exitCode=0 Nov 29 01:06:19 crc kubenswrapper[4931]: I1129 01:06:19.735875 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wn64l/crc-debug-hf6p2" event={"ID":"93130af3-0a61-4f95-8eba-36321928d128","Type":"ContainerDied","Data":"d2d540418c4d2e80a6bd5cf4b37fd296c7ac883630b18a876c663e2b4e6517c4"} Nov 29 01:06:19 crc kubenswrapper[4931]: I1129 01:06:19.735916 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wn64l/crc-debug-hf6p2" event={"ID":"93130af3-0a61-4f95-8eba-36321928d128","Type":"ContainerStarted","Data":"a3a181267364ae915024df0a68178f81bcb63f56182cb5cb6a5e1333afae1604"} Nov 29 01:06:19 crc kubenswrapper[4931]: I1129 01:06:19.802119 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wn64l/crc-debug-hf6p2"] Nov 29 01:06:19 crc kubenswrapper[4931]: I1129 01:06:19.813435 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wn64l/crc-debug-hf6p2"] Nov 29 01:06:20 crc kubenswrapper[4931]: I1129 01:06:20.213466 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:06:20 crc kubenswrapper[4931]: E1129 01:06:20.214033 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:06:20 crc kubenswrapper[4931]: I1129 01:06:20.861487 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/crc-debug-hf6p2" Nov 29 01:06:20 crc kubenswrapper[4931]: I1129 01:06:20.984764 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tb5k\" (UniqueName: \"kubernetes.io/projected/93130af3-0a61-4f95-8eba-36321928d128-kube-api-access-8tb5k\") pod \"93130af3-0a61-4f95-8eba-36321928d128\" (UID: \"93130af3-0a61-4f95-8eba-36321928d128\") " Nov 29 01:06:20 crc kubenswrapper[4931]: I1129 01:06:20.985094 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/93130af3-0a61-4f95-8eba-36321928d128-host\") pod \"93130af3-0a61-4f95-8eba-36321928d128\" (UID: \"93130af3-0a61-4f95-8eba-36321928d128\") " Nov 29 01:06:20 crc kubenswrapper[4931]: I1129 01:06:20.985232 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93130af3-0a61-4f95-8eba-36321928d128-host" (OuterVolumeSpecName: "host") pod "93130af3-0a61-4f95-8eba-36321928d128" (UID: "93130af3-0a61-4f95-8eba-36321928d128"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 01:06:20 crc kubenswrapper[4931]: I1129 01:06:20.985460 4931 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/93130af3-0a61-4f95-8eba-36321928d128-host\") on node \"crc\" DevicePath \"\"" Nov 29 01:06:20 crc kubenswrapper[4931]: I1129 01:06:20.996292 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93130af3-0a61-4f95-8eba-36321928d128-kube-api-access-8tb5k" (OuterVolumeSpecName: "kube-api-access-8tb5k") pod "93130af3-0a61-4f95-8eba-36321928d128" (UID: "93130af3-0a61-4f95-8eba-36321928d128"). InnerVolumeSpecName "kube-api-access-8tb5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:06:21 crc kubenswrapper[4931]: I1129 01:06:21.087113 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tb5k\" (UniqueName: \"kubernetes.io/projected/93130af3-0a61-4f95-8eba-36321928d128-kube-api-access-8tb5k\") on node \"crc\" DevicePath \"\"" Nov 29 01:06:21 crc kubenswrapper[4931]: I1129 01:06:21.223298 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93130af3-0a61-4f95-8eba-36321928d128" path="/var/lib/kubelet/pods/93130af3-0a61-4f95-8eba-36321928d128/volumes" Nov 29 01:06:21 crc kubenswrapper[4931]: I1129 01:06:21.754068 4931 scope.go:117] "RemoveContainer" containerID="d2d540418c4d2e80a6bd5cf4b37fd296c7ac883630b18a876c663e2b4e6517c4" Nov 29 01:06:21 crc kubenswrapper[4931]: I1129 01:06:21.754176 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/crc-debug-hf6p2" Nov 29 01:06:33 crc kubenswrapper[4931]: I1129 01:06:33.214380 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:06:33 crc kubenswrapper[4931]: E1129 01:06:33.215473 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:06:36 crc kubenswrapper[4931]: I1129 01:06:36.987836 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d5fd9f77d-c4gvt_d6aa6c84-7c6c-488e-bf3b-7393ad629ed4/barbican-api/0.log" Nov 29 01:06:36 crc kubenswrapper[4931]: I1129 01:06:36.998170 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d5fd9f77d-c4gvt_d6aa6c84-7c6c-488e-bf3b-7393ad629ed4/barbican-api-log/0.log" Nov 29 01:06:37 crc kubenswrapper[4931]: I1129 01:06:37.137107 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b65859cc8-zkpcj_5941c09d-aeac-449b-bd1f-b8d590ab5ecc/barbican-keystone-listener/0.log" Nov 29 01:06:37 crc kubenswrapper[4931]: I1129 01:06:37.214900 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b65859cc8-zkpcj_5941c09d-aeac-449b-bd1f-b8d590ab5ecc/barbican-keystone-listener-log/0.log" Nov 29 01:06:37 crc kubenswrapper[4931]: I1129 01:06:37.296482 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5644b469ff-dnc8z_2ba68cea-6add-4e52-96ad-458f708a02e0/barbican-worker/0.log" Nov 29 01:06:37 crc kubenswrapper[4931]: I1129 01:06:37.377311 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5644b469ff-dnc8z_2ba68cea-6add-4e52-96ad-458f708a02e0/barbican-worker-log/0.log" Nov 29 01:06:37 crc kubenswrapper[4931]: I1129 01:06:37.450440 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv_6a0a53a8-99d7-4a99-bdf0-67f72b8cca71/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:37 crc kubenswrapper[4931]: I1129 01:06:37.614205 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ecafd71-ca29-4382-9d56-d1e7422d113b/ceilometer-central-agent/0.log" Nov 29 01:06:37 crc kubenswrapper[4931]: I1129 01:06:37.654790 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ecafd71-ca29-4382-9d56-d1e7422d113b/proxy-httpd/0.log" Nov 29 01:06:37 crc kubenswrapper[4931]: I1129 01:06:37.660509 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ecafd71-ca29-4382-9d56-d1e7422d113b/ceilometer-notification-agent/0.log" Nov 29 01:06:37 crc kubenswrapper[4931]: I1129 01:06:37.715172 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ecafd71-ca29-4382-9d56-d1e7422d113b/sg-core/0.log" Nov 29 01:06:37 crc kubenswrapper[4931]: I1129 01:06:37.850339 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_80c8c44d-d0c0-4d7a-9063-9915b1c4aea6/cinder-api/0.log" Nov 29 01:06:37 crc kubenswrapper[4931]: I1129 01:06:37.865945 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_80c8c44d-d0c0-4d7a-9063-9915b1c4aea6/cinder-api-log/0.log" Nov 29 01:06:38 crc kubenswrapper[4931]: I1129 01:06:38.018345 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_06e9c000-a994-4023-a6a1-eaf84b161005/cinder-scheduler/0.log" Nov 29 01:06:38 crc kubenswrapper[4931]: I1129 01:06:38.667544 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp_3d56a6b7-f516-4675-b2bc-50372feb503d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:38 crc kubenswrapper[4931]: I1129 01:06:38.708149 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_06e9c000-a994-4023-a6a1-eaf84b161005/probe/0.log" Nov 29 01:06:38 crc kubenswrapper[4931]: I1129 01:06:38.926339 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-brwbq_5da7ac90-6a60-4473-b8b8-5531d4c2ac68/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:38 crc kubenswrapper[4931]: I1129 01:06:38.961168 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-578c4b6ff9-7b25t_76e2de18-8995-446b-aaf0-427974e0334b/init/0.log" Nov 29 01:06:39 crc kubenswrapper[4931]: I1129 01:06:39.117274 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-578c4b6ff9-7b25t_76e2de18-8995-446b-aaf0-427974e0334b/init/0.log" Nov 29 01:06:39 crc kubenswrapper[4931]: I1129 01:06:39.162801 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6_935c6e60-45b7-4859-bbd0-b147fa600465/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:39 crc kubenswrapper[4931]: I1129 01:06:39.211113 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-578c4b6ff9-7b25t_76e2de18-8995-446b-aaf0-427974e0334b/dnsmasq-dns/0.log" Nov 29 01:06:39 crc kubenswrapper[4931]: I1129 01:06:39.397886 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ea6acd51-8e93-4fdd-9535-84c7e6d2cee6/glance-httpd/0.log" Nov 29 01:06:39 crc kubenswrapper[4931]: I1129 01:06:39.406692 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ea6acd51-8e93-4fdd-9535-84c7e6d2cee6/glance-log/0.log" Nov 29 01:06:39 crc kubenswrapper[4931]: I1129 01:06:39.579041 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_78acdbbf-4fab-4641-86e8-49f124b8247e/glance-log/0.log" Nov 29 01:06:39 crc kubenswrapper[4931]: I1129 01:06:39.588346 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_78acdbbf-4fab-4641-86e8-49f124b8247e/glance-httpd/0.log" Nov 29 01:06:39 crc kubenswrapper[4931]: I1129 01:06:39.738876 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-79d99cf788-wwqzd_4003b615-a779-45a5-b95b-913c80226f4a/horizon/0.log" Nov 29 01:06:39 crc kubenswrapper[4931]: I1129 01:06:39.910392 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7_88f61c01-c5da-4d84-a484-48f4e3b4e3e4/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:39 crc kubenswrapper[4931]: I1129 01:06:39.992375 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-kgz5j_cb09a671-9026-4aec-ba97-911979c3f0d7/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:40 crc kubenswrapper[4931]: I1129 01:06:40.081496 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-79d99cf788-wwqzd_4003b615-a779-45a5-b95b-913c80226f4a/horizon-log/0.log" Nov 29 01:06:40 crc kubenswrapper[4931]: I1129 01:06:40.178302 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29406301-hhnfj_55460ba8-a9d1-4ef2-9bcf-e08d863698fa/keystone-cron/0.log" Nov 29 01:06:40 crc kubenswrapper[4931]: I1129 01:06:40.253351 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-84d76c555c-shgd6_aff73224-6284-4ef1-ace3-06eb11e904b8/keystone-api/0.log" Nov 29 01:06:40 crc kubenswrapper[4931]: I1129 01:06:40.301151 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b3edf21b-40ad-40e1-98db-78f609a9b4bc/kube-state-metrics/0.log" Nov 29 01:06:40 crc kubenswrapper[4931]: I1129 01:06:40.445656 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-r5958_ca54c57b-1fac-46bf-aa71-830b5417fca0/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:40 crc kubenswrapper[4931]: I1129 01:06:40.780012 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bcd7d89b9-l4sjs_e1ff7fdf-846a-4772-b8d8-fee7b15eed58/neutron-httpd/0.log" Nov 29 01:06:40 crc kubenswrapper[4931]: I1129 01:06:40.789113 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bcd7d89b9-l4sjs_e1ff7fdf-846a-4772-b8d8-fee7b15eed58/neutron-api/0.log" Nov 29 01:06:40 crc kubenswrapper[4931]: I1129 01:06:40.789568 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52_054b5e53-28bc-4222-9f59-ad0ebd2cf767/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:41 crc kubenswrapper[4931]: I1129 01:06:41.237173 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_611072c6-340e-49fc-870e-8a1e56f2cc90/nova-api-log/0.log" Nov 29 01:06:41 crc kubenswrapper[4931]: I1129 01:06:41.500617 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_2dcc4f63-f88c-436a-96a6-cae43c016e0c/nova-cell0-conductor-conductor/0.log" Nov 29 01:06:41 crc kubenswrapper[4931]: I1129 01:06:41.583382 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ecb51737-365b-4866-9d46-3e3aa5729b38/nova-cell1-conductor-conductor/0.log" Nov 29 01:06:41 crc kubenswrapper[4931]: I1129 01:06:41.802848 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_611072c6-340e-49fc-870e-8a1e56f2cc90/nova-api-api/0.log" Nov 29 01:06:41 crc kubenswrapper[4931]: I1129 01:06:41.826871 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_82ad9af8-53d9-48fc-b251-3b580018e51d/nova-cell1-novncproxy-novncproxy/0.log" Nov 29 01:06:41 crc kubenswrapper[4931]: I1129 01:06:41.910547 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-srlj9_aa634449-4a1f-4827-a318-5502c395886e/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:42 crc kubenswrapper[4931]: I1129 01:06:42.152570 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_6440833b-24fe-4da5-afae-6111fc7e760e/nova-metadata-log/0.log" Nov 29 01:06:42 crc kubenswrapper[4931]: I1129 01:06:42.364947 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2b4d256-9b86-48e3-98d8-266271a338dd/mysql-bootstrap/0.log" Nov 29 01:06:42 crc kubenswrapper[4931]: I1129 01:06:42.471904 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2/nova-scheduler-scheduler/0.log" Nov 29 01:06:42 crc kubenswrapper[4931]: I1129 01:06:42.553754 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2b4d256-9b86-48e3-98d8-266271a338dd/mysql-bootstrap/0.log" Nov 29 01:06:42 crc kubenswrapper[4931]: I1129 01:06:42.651053 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2b4d256-9b86-48e3-98d8-266271a338dd/galera/0.log" Nov 29 01:06:42 crc kubenswrapper[4931]: I1129 01:06:42.772340 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ba2777f2-23db-46b8-8ccf-cb0813a39678/mysql-bootstrap/0.log" Nov 29 01:06:42 crc kubenswrapper[4931]: I1129 01:06:42.954379 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ba2777f2-23db-46b8-8ccf-cb0813a39678/mysql-bootstrap/0.log" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.015101 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ba2777f2-23db-46b8-8ccf-cb0813a39678/galera/0.log" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.136225 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_c9276f25-9d5a-4953-9145-4da299973ff4/openstackclient/0.log" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.220762 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-m9dkp_f9b49437-b579-4f61-82a3-57d81e77b0b7/openstack-network-exporter/0.log" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.320739 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_6440833b-24fe-4da5-afae-6111fc7e760e/nova-metadata-metadata/0.log" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.414035 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7s9fx_28c11b0b-3026-4f9f-8de8-71174fbd19d6/ovsdb-server-init/0.log" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.601002 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7s9fx_28c11b0b-3026-4f9f-8de8-71174fbd19d6/ovsdb-server-init/0.log" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.631654 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7s9fx_28c11b0b-3026-4f9f-8de8-71174fbd19d6/ovsdb-server/0.log" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.647324 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7s9fx_28c11b0b-3026-4f9f-8de8-71174fbd19d6/ovs-vswitchd/0.log" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.789981 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-trsph_a64bc6bd-8033-45c9-b2d7-9602aea09225/ovn-controller/0.log" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.862930 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6tsqg"] Nov 29 01:06:43 crc kubenswrapper[4931]: E1129 01:06:43.863547 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93130af3-0a61-4f95-8eba-36321928d128" containerName="container-00" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.863565 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="93130af3-0a61-4f95-8eba-36321928d128" containerName="container-00" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.863742 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="93130af3-0a61-4f95-8eba-36321928d128" containerName="container-00" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.865080 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.883724 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6tsqg"] Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.890630 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-4p9cv_7715e2f2-11a1-4702-86f0-84744ae25c5d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.985013 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/488a91e3-6158-4bae-9210-a3c04ddfc8ac-utilities\") pod \"certified-operators-6tsqg\" (UID: \"488a91e3-6158-4bae-9210-a3c04ddfc8ac\") " pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.985056 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7stmp\" (UniqueName: \"kubernetes.io/projected/488a91e3-6158-4bae-9210-a3c04ddfc8ac-kube-api-access-7stmp\") pod \"certified-operators-6tsqg\" (UID: \"488a91e3-6158-4bae-9210-a3c04ddfc8ac\") " pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:06:43 crc kubenswrapper[4931]: I1129 01:06:43.985077 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/488a91e3-6158-4bae-9210-a3c04ddfc8ac-catalog-content\") pod \"certified-operators-6tsqg\" (UID: \"488a91e3-6158-4bae-9210-a3c04ddfc8ac\") " pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.031444 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb/openstack-network-exporter/0.log" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.076715 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb/ovn-northd/0.log" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.088962 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/488a91e3-6158-4bae-9210-a3c04ddfc8ac-utilities\") pod \"certified-operators-6tsqg\" (UID: \"488a91e3-6158-4bae-9210-a3c04ddfc8ac\") " pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.089674 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7stmp\" (UniqueName: \"kubernetes.io/projected/488a91e3-6158-4bae-9210-a3c04ddfc8ac-kube-api-access-7stmp\") pod \"certified-operators-6tsqg\" (UID: \"488a91e3-6158-4bae-9210-a3c04ddfc8ac\") " pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.089443 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/488a91e3-6158-4bae-9210-a3c04ddfc8ac-utilities\") pod \"certified-operators-6tsqg\" (UID: \"488a91e3-6158-4bae-9210-a3c04ddfc8ac\") " pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.089703 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/488a91e3-6158-4bae-9210-a3c04ddfc8ac-catalog-content\") pod \"certified-operators-6tsqg\" (UID: \"488a91e3-6158-4bae-9210-a3c04ddfc8ac\") " pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.090041 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/488a91e3-6158-4bae-9210-a3c04ddfc8ac-catalog-content\") pod \"certified-operators-6tsqg\" (UID: \"488a91e3-6158-4bae-9210-a3c04ddfc8ac\") " pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.135112 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7stmp\" (UniqueName: \"kubernetes.io/projected/488a91e3-6158-4bae-9210-a3c04ddfc8ac-kube-api-access-7stmp\") pod \"certified-operators-6tsqg\" (UID: \"488a91e3-6158-4bae-9210-a3c04ddfc8ac\") " pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.180104 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_979903af-0c72-45fe-a3de-0dc52370bc57/openstack-network-exporter/0.log" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.202754 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.360183 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_979903af-0c72-45fe-a3de-0dc52370bc57/ovsdbserver-nb/0.log" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.671016 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_37afe9f6-664f-4f51-90a9-485e87a1db92/ovsdbserver-sb/0.log" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.733221 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6tsqg"] Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.869493 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_37afe9f6-664f-4f51-90a9-485e87a1db92/openstack-network-exporter/0.log" Nov 29 01:06:44 crc kubenswrapper[4931]: I1129 01:06:44.896968 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6b5869645b-zlntx_48538810-7ec7-4d21-9cf3-1c0ee151eee5/placement-api/0.log" Nov 29 01:06:45 crc kubenswrapper[4931]: I1129 01:06:45.028278 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6b5869645b-zlntx_48538810-7ec7-4d21-9cf3-1c0ee151eee5/placement-log/0.log" Nov 29 01:06:45 crc kubenswrapper[4931]: I1129 01:06:45.048916 4931 generic.go:334] "Generic (PLEG): container finished" podID="488a91e3-6158-4bae-9210-a3c04ddfc8ac" containerID="88cd9c9345046b1e686c58074263168680616585a10d86171b4cf0401a70bb00" exitCode=0 Nov 29 01:06:45 crc kubenswrapper[4931]: I1129 01:06:45.048977 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6tsqg" event={"ID":"488a91e3-6158-4bae-9210-a3c04ddfc8ac","Type":"ContainerDied","Data":"88cd9c9345046b1e686c58074263168680616585a10d86171b4cf0401a70bb00"} Nov 29 01:06:45 crc kubenswrapper[4931]: I1129 01:06:45.049008 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6tsqg" event={"ID":"488a91e3-6158-4bae-9210-a3c04ddfc8ac","Type":"ContainerStarted","Data":"22a10333ba36a6a30369e7d858a60da0929bc08e3333a43f8c2f8e025e6a6e0f"} Nov 29 01:06:45 crc kubenswrapper[4931]: I1129 01:06:45.089901 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_421b0e7c-7dcd-4e58-a2ee-445c18ec4b22/setup-container/0.log" Nov 29 01:06:45 crc kubenswrapper[4931]: I1129 01:06:45.658079 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f8ccbdb2-db21-4fce-9111-b34a6800e238/setup-container/0.log" Nov 29 01:06:45 crc kubenswrapper[4931]: I1129 01:06:45.661715 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_421b0e7c-7dcd-4e58-a2ee-445c18ec4b22/rabbitmq/0.log" Nov 29 01:06:45 crc kubenswrapper[4931]: I1129 01:06:45.760035 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_421b0e7c-7dcd-4e58-a2ee-445c18ec4b22/setup-container/0.log" Nov 29 01:06:45 crc kubenswrapper[4931]: I1129 01:06:45.971062 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f8ccbdb2-db21-4fce-9111-b34a6800e238/rabbitmq/0.log" Nov 29 01:06:46 crc kubenswrapper[4931]: I1129 01:06:46.045145 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f8ccbdb2-db21-4fce-9111-b34a6800e238/setup-container/0.log" Nov 29 01:06:46 crc kubenswrapper[4931]: I1129 01:06:46.070147 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs_417e828a-2ea5-4ad0-86e0-b5d612cccbe2/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:46 crc kubenswrapper[4931]: I1129 01:06:46.289380 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-n6rmv_2921aa97-795f-434a-ac6e-009e6555e864/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:46 crc kubenswrapper[4931]: I1129 01:06:46.298778 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9_8449a5b7-2fa6-4966-9d34-0243e3536f2b/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:46 crc kubenswrapper[4931]: I1129 01:06:46.514252 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-gfsvs_333e0546-64b0-412e-9059-1e7a4082d360/ssh-known-hosts-edpm-deployment/0.log" Nov 29 01:06:46 crc kubenswrapper[4931]: I1129 01:06:46.553783 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-9gcqp_6c31a7d4-ea5c-4434-a405-cd4148ca5a87/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:46 crc kubenswrapper[4931]: I1129 01:06:46.749734 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-764cb44577-7q67f_550e16f7-f101-4a91-8e2f-45af3430d6b9/proxy-server/0.log" Nov 29 01:06:46 crc kubenswrapper[4931]: I1129 01:06:46.841458 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-764cb44577-7q67f_550e16f7-f101-4a91-8e2f-45af3430d6b9/proxy-httpd/0.log" Nov 29 01:06:46 crc kubenswrapper[4931]: I1129 01:06:46.915609 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-jnxgv_64db2499-e411-42b9-8e99-92e748a9ad5d/swift-ring-rebalance/0.log" Nov 29 01:06:47 crc kubenswrapper[4931]: I1129 01:06:47.220587 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:06:47 crc kubenswrapper[4931]: E1129 01:06:47.224677 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:06:47 crc kubenswrapper[4931]: I1129 01:06:47.589100 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/account-auditor/0.log" Nov 29 01:06:47 crc kubenswrapper[4931]: I1129 01:06:47.595367 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/account-replicator/0.log" Nov 29 01:06:47 crc kubenswrapper[4931]: I1129 01:06:47.601364 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/account-reaper/0.log" Nov 29 01:06:47 crc kubenswrapper[4931]: I1129 01:06:47.678346 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/account-server/0.log" Nov 29 01:06:47 crc kubenswrapper[4931]: I1129 01:06:47.808131 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/container-auditor/0.log" Nov 29 01:06:47 crc kubenswrapper[4931]: I1129 01:06:47.814849 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/container-server/0.log" Nov 29 01:06:47 crc kubenswrapper[4931]: I1129 01:06:47.824547 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/container-replicator/0.log" Nov 29 01:06:47 crc kubenswrapper[4931]: I1129 01:06:47.882352 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/container-updater/0.log" Nov 29 01:06:48 crc kubenswrapper[4931]: I1129 01:06:48.023112 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/object-expirer/0.log" Nov 29 01:06:48 crc kubenswrapper[4931]: I1129 01:06:48.027999 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/object-auditor/0.log" Nov 29 01:06:48 crc kubenswrapper[4931]: I1129 01:06:48.078801 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/object-server/0.log" Nov 29 01:06:48 crc kubenswrapper[4931]: I1129 01:06:48.080918 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/object-replicator/0.log" Nov 29 01:06:48 crc kubenswrapper[4931]: I1129 01:06:48.237923 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/object-updater/0.log" Nov 29 01:06:48 crc kubenswrapper[4931]: I1129 01:06:48.286129 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/swift-recon-cron/0.log" Nov 29 01:06:48 crc kubenswrapper[4931]: I1129 01:06:48.321245 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/rsync/0.log" Nov 29 01:06:48 crc kubenswrapper[4931]: I1129 01:06:48.492095 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj_6a410480-c08c-4a1a-b3a4-6d2372d6bb2a/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:48 crc kubenswrapper[4931]: I1129 01:06:48.572615 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_16c927ea-8729-4da8-8893-c23a53d85c5d/tempest-tests-tempest-tests-runner/0.log" Nov 29 01:06:48 crc kubenswrapper[4931]: I1129 01:06:48.675366 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_dbcaae8b-9df7-4dda-a4ea-3a237c20627f/test-operator-logs-container/0.log" Nov 29 01:06:48 crc kubenswrapper[4931]: I1129 01:06:48.758789 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-g26zs_01b89e61-27d6-407a-97e0-7ced1cf90139/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:06:55 crc kubenswrapper[4931]: I1129 01:06:55.148862 4931 generic.go:334] "Generic (PLEG): container finished" podID="488a91e3-6158-4bae-9210-a3c04ddfc8ac" containerID="1ee31d2af189b2fd4f30fec0552ed2faae371edeab224b0bb08411315386c53c" exitCode=0 Nov 29 01:06:55 crc kubenswrapper[4931]: I1129 01:06:55.149369 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6tsqg" event={"ID":"488a91e3-6158-4bae-9210-a3c04ddfc8ac","Type":"ContainerDied","Data":"1ee31d2af189b2fd4f30fec0552ed2faae371edeab224b0bb08411315386c53c"} Nov 29 01:06:57 crc kubenswrapper[4931]: I1129 01:06:57.156585 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_d60ec023-b0d6-466c-b188-5e0e2b47a962/memcached/0.log" Nov 29 01:06:57 crc kubenswrapper[4931]: I1129 01:06:57.167825 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6tsqg" event={"ID":"488a91e3-6158-4bae-9210-a3c04ddfc8ac","Type":"ContainerStarted","Data":"5364beba7708e01b921f433ef367ce0ca463422d46314cfdc523da6e6094dceb"} Nov 29 01:06:57 crc kubenswrapper[4931]: I1129 01:06:57.189769 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6tsqg" podStartSLOduration=3.192356677 podStartE2EDuration="14.189736583s" podCreationTimestamp="2025-11-29 01:06:43 +0000 UTC" firstStartedPulling="2025-11-29 01:06:45.051188338 +0000 UTC m=+4018.213081570" lastFinishedPulling="2025-11-29 01:06:56.048568244 +0000 UTC m=+4029.210461476" observedRunningTime="2025-11-29 01:06:57.186235624 +0000 UTC m=+4030.348128866" watchObservedRunningTime="2025-11-29 01:06:57.189736583 +0000 UTC m=+4030.351629835" Nov 29 01:06:59 crc kubenswrapper[4931]: I1129 01:06:59.212425 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:06:59 crc kubenswrapper[4931]: E1129 01:06:59.212892 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:07:04 crc kubenswrapper[4931]: I1129 01:07:04.204301 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:07:04 crc kubenswrapper[4931]: I1129 01:07:04.204768 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:07:04 crc kubenswrapper[4931]: I1129 01:07:04.257906 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:07:04 crc kubenswrapper[4931]: I1129 01:07:04.312342 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6tsqg" Nov 29 01:07:04 crc kubenswrapper[4931]: I1129 01:07:04.390562 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6tsqg"] Nov 29 01:07:04 crc kubenswrapper[4931]: I1129 01:07:04.502065 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2jpnh"] Nov 29 01:07:04 crc kubenswrapper[4931]: I1129 01:07:04.502276 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2jpnh" podUID="59f58c30-f7fe-4141-ab70-122e133cccdc" containerName="registry-server" containerID="cri-o://82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf" gracePeriod=2 Nov 29 01:07:04 crc kubenswrapper[4931]: I1129 01:07:04.958664 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.059502 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59f58c30-f7fe-4141-ab70-122e133cccdc-catalog-content\") pod \"59f58c30-f7fe-4141-ab70-122e133cccdc\" (UID: \"59f58c30-f7fe-4141-ab70-122e133cccdc\") " Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.059802 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75vcr\" (UniqueName: \"kubernetes.io/projected/59f58c30-f7fe-4141-ab70-122e133cccdc-kube-api-access-75vcr\") pod \"59f58c30-f7fe-4141-ab70-122e133cccdc\" (UID: \"59f58c30-f7fe-4141-ab70-122e133cccdc\") " Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.059881 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59f58c30-f7fe-4141-ab70-122e133cccdc-utilities\") pod \"59f58c30-f7fe-4141-ab70-122e133cccdc\" (UID: \"59f58c30-f7fe-4141-ab70-122e133cccdc\") " Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.060918 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f58c30-f7fe-4141-ab70-122e133cccdc-utilities" (OuterVolumeSpecName: "utilities") pod "59f58c30-f7fe-4141-ab70-122e133cccdc" (UID: "59f58c30-f7fe-4141-ab70-122e133cccdc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.069668 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59f58c30-f7fe-4141-ab70-122e133cccdc-kube-api-access-75vcr" (OuterVolumeSpecName: "kube-api-access-75vcr") pod "59f58c30-f7fe-4141-ab70-122e133cccdc" (UID: "59f58c30-f7fe-4141-ab70-122e133cccdc"). InnerVolumeSpecName "kube-api-access-75vcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.130531 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f58c30-f7fe-4141-ab70-122e133cccdc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59f58c30-f7fe-4141-ab70-122e133cccdc" (UID: "59f58c30-f7fe-4141-ab70-122e133cccdc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.162346 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59f58c30-f7fe-4141-ab70-122e133cccdc-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.162378 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59f58c30-f7fe-4141-ab70-122e133cccdc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.162389 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75vcr\" (UniqueName: \"kubernetes.io/projected/59f58c30-f7fe-4141-ab70-122e133cccdc-kube-api-access-75vcr\") on node \"crc\" DevicePath \"\"" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.233557 4931 generic.go:334] "Generic (PLEG): container finished" podID="59f58c30-f7fe-4141-ab70-122e133cccdc" containerID="82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf" exitCode=0 Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.233605 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2jpnh" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.233643 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jpnh" event={"ID":"59f58c30-f7fe-4141-ab70-122e133cccdc","Type":"ContainerDied","Data":"82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf"} Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.233698 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jpnh" event={"ID":"59f58c30-f7fe-4141-ab70-122e133cccdc","Type":"ContainerDied","Data":"7a49c0df9d2e09ff7322089055954337d821b48ddc4ed2dee825d55e5055d1f5"} Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.233717 4931 scope.go:117] "RemoveContainer" containerID="82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.256983 4931 scope.go:117] "RemoveContainer" containerID="1d4dadae7ba46d919333c87e686f8afeedf76bc6890d51ebaf31f80e1a91c16a" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.265201 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2jpnh"] Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.277363 4931 scope.go:117] "RemoveContainer" containerID="7f42064d17eef14c72ba5fa14d986fe9d400d78de8d8cfff0c65bcc45530dedd" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.282732 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2jpnh"] Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.319517 4931 scope.go:117] "RemoveContainer" containerID="82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf" Nov 29 01:07:05 crc kubenswrapper[4931]: E1129 01:07:05.319913 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf\": container with ID starting with 82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf not found: ID does not exist" containerID="82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.319941 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf"} err="failed to get container status \"82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf\": rpc error: code = NotFound desc = could not find container \"82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf\": container with ID starting with 82505710ff91e8e6c742aa45c2842b74e92495f065bccac0da77c98270322bbf not found: ID does not exist" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.319961 4931 scope.go:117] "RemoveContainer" containerID="1d4dadae7ba46d919333c87e686f8afeedf76bc6890d51ebaf31f80e1a91c16a" Nov 29 01:07:05 crc kubenswrapper[4931]: E1129 01:07:05.320192 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d4dadae7ba46d919333c87e686f8afeedf76bc6890d51ebaf31f80e1a91c16a\": container with ID starting with 1d4dadae7ba46d919333c87e686f8afeedf76bc6890d51ebaf31f80e1a91c16a not found: ID does not exist" containerID="1d4dadae7ba46d919333c87e686f8afeedf76bc6890d51ebaf31f80e1a91c16a" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.320214 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d4dadae7ba46d919333c87e686f8afeedf76bc6890d51ebaf31f80e1a91c16a"} err="failed to get container status \"1d4dadae7ba46d919333c87e686f8afeedf76bc6890d51ebaf31f80e1a91c16a\": rpc error: code = NotFound desc = could not find container \"1d4dadae7ba46d919333c87e686f8afeedf76bc6890d51ebaf31f80e1a91c16a\": container with ID starting with 1d4dadae7ba46d919333c87e686f8afeedf76bc6890d51ebaf31f80e1a91c16a not found: ID does not exist" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.320228 4931 scope.go:117] "RemoveContainer" containerID="7f42064d17eef14c72ba5fa14d986fe9d400d78de8d8cfff0c65bcc45530dedd" Nov 29 01:07:05 crc kubenswrapper[4931]: E1129 01:07:05.320523 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f42064d17eef14c72ba5fa14d986fe9d400d78de8d8cfff0c65bcc45530dedd\": container with ID starting with 7f42064d17eef14c72ba5fa14d986fe9d400d78de8d8cfff0c65bcc45530dedd not found: ID does not exist" containerID="7f42064d17eef14c72ba5fa14d986fe9d400d78de8d8cfff0c65bcc45530dedd" Nov 29 01:07:05 crc kubenswrapper[4931]: I1129 01:07:05.320543 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f42064d17eef14c72ba5fa14d986fe9d400d78de8d8cfff0c65bcc45530dedd"} err="failed to get container status \"7f42064d17eef14c72ba5fa14d986fe9d400d78de8d8cfff0c65bcc45530dedd\": rpc error: code = NotFound desc = could not find container \"7f42064d17eef14c72ba5fa14d986fe9d400d78de8d8cfff0c65bcc45530dedd\": container with ID starting with 7f42064d17eef14c72ba5fa14d986fe9d400d78de8d8cfff0c65bcc45530dedd not found: ID does not exist" Nov 29 01:07:07 crc kubenswrapper[4931]: I1129 01:07:07.223911 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59f58c30-f7fe-4141-ab70-122e133cccdc" path="/var/lib/kubelet/pods/59f58c30-f7fe-4141-ab70-122e133cccdc/volumes" Nov 29 01:07:12 crc kubenswrapper[4931]: I1129 01:07:12.213138 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:07:12 crc kubenswrapper[4931]: E1129 01:07:12.214063 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:07:16 crc kubenswrapper[4931]: I1129 01:07:16.057928 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/util/0.log" Nov 29 01:07:16 crc kubenswrapper[4931]: I1129 01:07:16.278887 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/pull/0.log" Nov 29 01:07:16 crc kubenswrapper[4931]: I1129 01:07:16.312976 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/util/0.log" Nov 29 01:07:16 crc kubenswrapper[4931]: I1129 01:07:16.322621 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/pull/0.log" Nov 29 01:07:16 crc kubenswrapper[4931]: I1129 01:07:16.497463 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/pull/0.log" Nov 29 01:07:16 crc kubenswrapper[4931]: I1129 01:07:16.519922 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/util/0.log" Nov 29 01:07:16 crc kubenswrapper[4931]: I1129 01:07:16.525201 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/extract/0.log" Nov 29 01:07:16 crc kubenswrapper[4931]: I1129 01:07:16.646544 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-qkqqf_64f054e4-1ff5-4d20-a098-499b8baa25eb/kube-rbac-proxy/0.log" Nov 29 01:07:16 crc kubenswrapper[4931]: I1129 01:07:16.749405 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-qkqqf_64f054e4-1ff5-4d20-a098-499b8baa25eb/manager/0.log" Nov 29 01:07:16 crc kubenswrapper[4931]: I1129 01:07:16.803913 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-tqw4f_551c1c83-9dd7-47c7-8411-30881adc79bb/kube-rbac-proxy/0.log" Nov 29 01:07:16 crc kubenswrapper[4931]: I1129 01:07:16.889054 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-tqw4f_551c1c83-9dd7-47c7-8411-30881adc79bb/manager/0.log" Nov 29 01:07:16 crc kubenswrapper[4931]: I1129 01:07:16.972125 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-wzvc6_256e71e1-5f65-4842-883e-1e3178e91965/kube-rbac-proxy/0.log" Nov 29 01:07:17 crc kubenswrapper[4931]: I1129 01:07:17.023598 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-wzvc6_256e71e1-5f65-4842-883e-1e3178e91965/manager/0.log" Nov 29 01:07:17 crc kubenswrapper[4931]: I1129 01:07:17.126841 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-lhc87_b870bbdd-21a8-4279-9d2a-12dd26745325/kube-rbac-proxy/0.log" Nov 29 01:07:17 crc kubenswrapper[4931]: I1129 01:07:17.313043 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-lhc87_b870bbdd-21a8-4279-9d2a-12dd26745325/manager/0.log" Nov 29 01:07:17 crc kubenswrapper[4931]: I1129 01:07:17.444220 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-nmtz7_21842304-a783-4f75-bbb5-2ece9dacb6ca/kube-rbac-proxy/0.log" Nov 29 01:07:17 crc kubenswrapper[4931]: I1129 01:07:17.445265 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-nmtz7_21842304-a783-4f75-bbb5-2ece9dacb6ca/manager/0.log" Nov 29 01:07:17 crc kubenswrapper[4931]: I1129 01:07:17.555251 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-r5d5m_6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c/kube-rbac-proxy/0.log" Nov 29 01:07:17 crc kubenswrapper[4931]: I1129 01:07:17.643133 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-r5d5m_6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c/manager/0.log" Nov 29 01:07:17 crc kubenswrapper[4931]: I1129 01:07:17.690216 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-nkmhr_3a157886-d4b5-4019-a0d6-0b40f6509b21/kube-rbac-proxy/0.log" Nov 29 01:07:17 crc kubenswrapper[4931]: I1129 01:07:17.874155 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-zb94f_5f790baf-d77a-41dd-840c-f4d9e13420e1/kube-rbac-proxy/0.log" Nov 29 01:07:17 crc kubenswrapper[4931]: I1129 01:07:17.908373 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-nkmhr_3a157886-d4b5-4019-a0d6-0b40f6509b21/manager/0.log" Nov 29 01:07:17 crc kubenswrapper[4931]: I1129 01:07:17.919366 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-zb94f_5f790baf-d77a-41dd-840c-f4d9e13420e1/manager/0.log" Nov 29 01:07:18 crc kubenswrapper[4931]: I1129 01:07:18.058672 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-g82st_0ce6e577-25c0-4aee-b14f-47e02ebfbef9/kube-rbac-proxy/0.log" Nov 29 01:07:18 crc kubenswrapper[4931]: I1129 01:07:18.182610 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-g82st_0ce6e577-25c0-4aee-b14f-47e02ebfbef9/manager/0.log" Nov 29 01:07:18 crc kubenswrapper[4931]: I1129 01:07:18.258482 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-b4skl_5a12e787-bf00-4f2b-8760-9b9a5cad0834/kube-rbac-proxy/0.log" Nov 29 01:07:18 crc kubenswrapper[4931]: I1129 01:07:18.325070 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-b4skl_5a12e787-bf00-4f2b-8760-9b9a5cad0834/manager/0.log" Nov 29 01:07:18 crc kubenswrapper[4931]: I1129 01:07:18.368535 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-2dbgl_724defda-f0a4-4edd-8a1a-7f6625893fe7/kube-rbac-proxy/0.log" Nov 29 01:07:18 crc kubenswrapper[4931]: I1129 01:07:18.448164 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-2dbgl_724defda-f0a4-4edd-8a1a-7f6625893fe7/manager/0.log" Nov 29 01:07:18 crc kubenswrapper[4931]: I1129 01:07:18.558571 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-q7q7q_365eb479-0132-4a0c-99d3-01cb923a9d61/kube-rbac-proxy/0.log" Nov 29 01:07:18 crc kubenswrapper[4931]: I1129 01:07:18.594189 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-q7q7q_365eb479-0132-4a0c-99d3-01cb923a9d61/manager/0.log" Nov 29 01:07:18 crc kubenswrapper[4931]: I1129 01:07:18.703600 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-lpf99_c352f7a8-72dc-41da-9673-7fccd1a0974f/kube-rbac-proxy/0.log" Nov 29 01:07:18 crc kubenswrapper[4931]: I1129 01:07:18.817277 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-lpf99_c352f7a8-72dc-41da-9673-7fccd1a0974f/manager/0.log" Nov 29 01:07:18 crc kubenswrapper[4931]: I1129 01:07:18.889149 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-z2mjc_4c70c5cd-a329-4c90-b454-f836ddaef38f/kube-rbac-proxy/0.log" Nov 29 01:07:18 crc kubenswrapper[4931]: I1129 01:07:18.942748 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-z2mjc_4c70c5cd-a329-4c90-b454-f836ddaef38f/manager/0.log" Nov 29 01:07:19 crc kubenswrapper[4931]: I1129 01:07:19.048463 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp_fd6606d5-8cd5-4308-a010-eea75cb666f1/kube-rbac-proxy/0.log" Nov 29 01:07:19 crc kubenswrapper[4931]: I1129 01:07:19.104942 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp_fd6606d5-8cd5-4308-a010-eea75cb666f1/manager/0.log" Nov 29 01:07:19 crc kubenswrapper[4931]: I1129 01:07:19.365713 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-h7scz_4ef2db4a-ba6d-47a1-8054-b61f854da2d0/registry-server/0.log" Nov 29 01:07:19 crc kubenswrapper[4931]: I1129 01:07:19.438320 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5757985b9c-rwxqq_fe758412-e16a-4c07-8c0d-e79e9f54cd73/operator/0.log" Nov 29 01:07:19 crc kubenswrapper[4931]: I1129 01:07:19.642925 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lnrrq_6f4cd8c8-caf0-4191-a3c4-51a172812b9b/kube-rbac-proxy/0.log" Nov 29 01:07:19 crc kubenswrapper[4931]: I1129 01:07:19.772240 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lnrrq_6f4cd8c8-caf0-4191-a3c4-51a172812b9b/manager/0.log" Nov 29 01:07:19 crc kubenswrapper[4931]: I1129 01:07:19.851833 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-lc47s_96c0b73e-c18e-4a52-a20c-e10db096b799/manager/0.log" Nov 29 01:07:19 crc kubenswrapper[4931]: I1129 01:07:19.858832 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-lc47s_96c0b73e-c18e-4a52-a20c-e10db096b799/kube-rbac-proxy/0.log" Nov 29 01:07:20 crc kubenswrapper[4931]: I1129 01:07:20.051163 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-4qjb4_22500bd3-4958-4ba1-9b59-deb3032aea18/operator/0.log" Nov 29 01:07:20 crc kubenswrapper[4931]: I1129 01:07:20.082834 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-f62b4_249844ae-a64c-4f9d-b204-bf3a1956aaa7/kube-rbac-proxy/0.log" Nov 29 01:07:20 crc kubenswrapper[4931]: I1129 01:07:20.278050 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-lthd8_b57413c5-3888-496c-ad0b-18128019b945/kube-rbac-proxy/0.log" Nov 29 01:07:20 crc kubenswrapper[4931]: I1129 01:07:20.280376 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-f62b4_249844ae-a64c-4f9d-b204-bf3a1956aaa7/manager/0.log" Nov 29 01:07:20 crc kubenswrapper[4931]: I1129 01:07:20.377397 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-lthd8_b57413c5-3888-496c-ad0b-18128019b945/manager/0.log" Nov 29 01:07:20 crc kubenswrapper[4931]: I1129 01:07:20.382637 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74fb9455b7-sz2k9_0c61786d-9cf4-4447-82f7-47cc17ac65d8/manager/0.log" Nov 29 01:07:20 crc kubenswrapper[4931]: I1129 01:07:20.453232 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-dcl69_9621ec40-d523-423e-9bbe-4b063355cf08/kube-rbac-proxy/0.log" Nov 29 01:07:20 crc kubenswrapper[4931]: I1129 01:07:20.505773 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-dcl69_9621ec40-d523-423e-9bbe-4b063355cf08/manager/0.log" Nov 29 01:07:20 crc kubenswrapper[4931]: I1129 01:07:20.585558 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-qp9ht_2792db7c-cdd1-4356-97d5-fdba06bd44e4/kube-rbac-proxy/0.log" Nov 29 01:07:20 crc kubenswrapper[4931]: I1129 01:07:20.668490 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-qp9ht_2792db7c-cdd1-4356-97d5-fdba06bd44e4/manager/0.log" Nov 29 01:07:25 crc kubenswrapper[4931]: I1129 01:07:25.215547 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:07:25 crc kubenswrapper[4931]: E1129 01:07:25.216321 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:07:38 crc kubenswrapper[4931]: I1129 01:07:38.215358 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:07:38 crc kubenswrapper[4931]: E1129 01:07:38.216597 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:07:41 crc kubenswrapper[4931]: I1129 01:07:41.858243 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-pgslt_b6c59839-1a09-4b0b-90ac-eb9d001e9581/control-plane-machine-set-operator/0.log" Nov 29 01:07:42 crc kubenswrapper[4931]: I1129 01:07:42.017069 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xl2bw_980879e9-818d-414c-97c0-ad63e6cf4039/kube-rbac-proxy/0.log" Nov 29 01:07:42 crc kubenswrapper[4931]: I1129 01:07:42.049529 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xl2bw_980879e9-818d-414c-97c0-ad63e6cf4039/machine-api-operator/0.log" Nov 29 01:07:52 crc kubenswrapper[4931]: I1129 01:07:52.211789 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:07:52 crc kubenswrapper[4931]: E1129 01:07:52.212463 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:07:57 crc kubenswrapper[4931]: I1129 01:07:57.476052 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-s9fhh_ea00643b-3840-4675-b14a-323b377e2339/cert-manager-controller/0.log" Nov 29 01:07:57 crc kubenswrapper[4931]: I1129 01:07:57.579325 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-g6bkx_2f242144-c19c-4886-a51b-8547f684e2e9/cert-manager-cainjector/0.log" Nov 29 01:07:57 crc kubenswrapper[4931]: I1129 01:07:57.625525 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-b8cd4_283e1d51-cb76-418b-b4b0-d444320900ff/cert-manager-webhook/0.log" Nov 29 01:08:05 crc kubenswrapper[4931]: I1129 01:08:05.212945 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:08:05 crc kubenswrapper[4931]: E1129 01:08:05.213729 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:08:12 crc kubenswrapper[4931]: I1129 01:08:12.097500 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-g8dpm_6fbf3473-ee9a-47b5-a782-47948ff5e590/nmstate-console-plugin/0.log" Nov 29 01:08:12 crc kubenswrapper[4931]: I1129 01:08:12.294237 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-wv5fd_7622d8c6-7baa-47e7-8f03-52ccd0d2f93f/nmstate-handler/0.log" Nov 29 01:08:12 crc kubenswrapper[4931]: I1129 01:08:12.341290 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-cmcf4_e84ff97d-2d9a-4a4a-b619-a26e9c5a5595/kube-rbac-proxy/0.log" Nov 29 01:08:12 crc kubenswrapper[4931]: I1129 01:08:12.346198 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-cmcf4_e84ff97d-2d9a-4a4a-b619-a26e9c5a5595/nmstate-metrics/0.log" Nov 29 01:08:12 crc kubenswrapper[4931]: I1129 01:08:12.502156 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-t6vpl_3f3c7035-b1e3-48bf-bbb9-3c6400ffedae/nmstate-operator/0.log" Nov 29 01:08:12 crc kubenswrapper[4931]: I1129 01:08:12.550907 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-jh9h6_4d93b5c8-a612-45f9-9351-31ffc5eb3b5a/nmstate-webhook/0.log" Nov 29 01:08:18 crc kubenswrapper[4931]: I1129 01:08:18.212159 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:08:18 crc kubenswrapper[4931]: E1129 01:08:18.212967 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:08:29 crc kubenswrapper[4931]: I1129 01:08:29.149189 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-qzb8h_a44839c5-bede-4c88-b3ff-75f0d06074a6/kube-rbac-proxy/0.log" Nov 29 01:08:29 crc kubenswrapper[4931]: I1129 01:08:29.263891 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-qzb8h_a44839c5-bede-4c88-b3ff-75f0d06074a6/controller/0.log" Nov 29 01:08:29 crc kubenswrapper[4931]: I1129 01:08:29.396493 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-frr-files/0.log" Nov 29 01:08:29 crc kubenswrapper[4931]: I1129 01:08:29.546771 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-reloader/0.log" Nov 29 01:08:29 crc kubenswrapper[4931]: I1129 01:08:29.547534 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-frr-files/0.log" Nov 29 01:08:29 crc kubenswrapper[4931]: I1129 01:08:29.581050 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-metrics/0.log" Nov 29 01:08:29 crc kubenswrapper[4931]: I1129 01:08:29.594521 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-reloader/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.230141 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-reloader/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.238790 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-metrics/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.239708 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-frr-files/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.288864 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-metrics/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.462255 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-frr-files/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.494867 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-metrics/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.500776 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-reloader/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.517364 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/controller/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.640890 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/frr-metrics/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.673689 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/kube-rbac-proxy/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.727689 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/kube-rbac-proxy-frr/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.835609 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/reloader/0.log" Nov 29 01:08:30 crc kubenswrapper[4931]: I1129 01:08:30.919681 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-c96lv_fe08c847-a4ed-4e65-a3cb-e2145c69365f/frr-k8s-webhook-server/0.log" Nov 29 01:08:31 crc kubenswrapper[4931]: I1129 01:08:31.164096 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-df4f9f548-5vp72_70a74ef6-a92b-49b3-b533-17b6a15f513c/manager/0.log" Nov 29 01:08:31 crc kubenswrapper[4931]: I1129 01:08:31.262147 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-d7b5f7bd5-t4z72_04d440db-7f7e-46fa-9654-58b389ea3ad7/webhook-server/0.log" Nov 29 01:08:31 crc kubenswrapper[4931]: I1129 01:08:31.430858 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6ck62_2eb6f229-405c-40db-8504-8fb851cb98d4/kube-rbac-proxy/0.log" Nov 29 01:08:31 crc kubenswrapper[4931]: I1129 01:08:31.822845 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6ck62_2eb6f229-405c-40db-8504-8fb851cb98d4/speaker/0.log" Nov 29 01:08:31 crc kubenswrapper[4931]: I1129 01:08:31.905600 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/frr/0.log" Nov 29 01:08:32 crc kubenswrapper[4931]: I1129 01:08:32.212529 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:08:32 crc kubenswrapper[4931]: E1129 01:08:32.212988 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:08:44 crc kubenswrapper[4931]: I1129 01:08:44.212514 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:08:44 crc kubenswrapper[4931]: E1129 01:08:44.213155 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:08:45 crc kubenswrapper[4931]: I1129 01:08:45.880527 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/util/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.078565 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/pull/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.133116 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/pull/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.155859 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/util/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.260376 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/pull/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.269457 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/util/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.307695 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/extract/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.442190 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/util/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.571850 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/util/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.602518 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/pull/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.613071 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/pull/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.789755 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/pull/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.801209 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/util/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.806365 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/extract/0.log" Nov 29 01:08:46 crc kubenswrapper[4931]: I1129 01:08:46.999894 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/extract-utilities/0.log" Nov 29 01:08:47 crc kubenswrapper[4931]: I1129 01:08:47.208505 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/extract-utilities/0.log" Nov 29 01:08:47 crc kubenswrapper[4931]: I1129 01:08:47.223688 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/extract-content/0.log" Nov 29 01:08:47 crc kubenswrapper[4931]: I1129 01:08:47.244652 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/extract-content/0.log" Nov 29 01:08:47 crc kubenswrapper[4931]: I1129 01:08:47.437717 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/extract-utilities/0.log" Nov 29 01:08:47 crc kubenswrapper[4931]: I1129 01:08:47.479140 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/extract-content/0.log" Nov 29 01:08:47 crc kubenswrapper[4931]: I1129 01:08:47.616063 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/registry-server/0.log" Nov 29 01:08:47 crc kubenswrapper[4931]: I1129 01:08:47.660011 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zn6ld_c92e6619-370d-46ea-8925-a35dff600120/extract-utilities/0.log" Nov 29 01:08:47 crc kubenswrapper[4931]: I1129 01:08:47.816830 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zn6ld_c92e6619-370d-46ea-8925-a35dff600120/extract-utilities/0.log" Nov 29 01:08:47 crc kubenswrapper[4931]: I1129 01:08:47.820622 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zn6ld_c92e6619-370d-46ea-8925-a35dff600120/extract-content/0.log" Nov 29 01:08:47 crc kubenswrapper[4931]: I1129 01:08:47.889573 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zn6ld_c92e6619-370d-46ea-8925-a35dff600120/extract-content/0.log" Nov 29 01:08:48 crc kubenswrapper[4931]: I1129 01:08:48.043518 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zn6ld_c92e6619-370d-46ea-8925-a35dff600120/extract-utilities/0.log" Nov 29 01:08:48 crc kubenswrapper[4931]: I1129 01:08:48.090655 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zn6ld_c92e6619-370d-46ea-8925-a35dff600120/extract-content/0.log" Nov 29 01:08:48 crc kubenswrapper[4931]: I1129 01:08:48.239268 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-m8pmr_db3078a6-11fa-4bda-8b8e-d3ac07d1db11/marketplace-operator/0.log" Nov 29 01:08:48 crc kubenswrapper[4931]: I1129 01:08:48.368949 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/extract-utilities/0.log" Nov 29 01:08:48 crc kubenswrapper[4931]: I1129 01:08:48.592988 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/extract-utilities/0.log" Nov 29 01:08:48 crc kubenswrapper[4931]: I1129 01:08:48.607136 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/extract-content/0.log" Nov 29 01:08:48 crc kubenswrapper[4931]: I1129 01:08:48.737391 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zn6ld_c92e6619-370d-46ea-8925-a35dff600120/registry-server/0.log" Nov 29 01:08:48 crc kubenswrapper[4931]: I1129 01:08:48.760999 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/extract-content/0.log" Nov 29 01:08:48 crc kubenswrapper[4931]: I1129 01:08:48.926574 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/extract-content/0.log" Nov 29 01:08:48 crc kubenswrapper[4931]: I1129 01:08:48.965536 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/extract-utilities/0.log" Nov 29 01:08:49 crc kubenswrapper[4931]: I1129 01:08:49.387316 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/extract-utilities/0.log" Nov 29 01:08:49 crc kubenswrapper[4931]: I1129 01:08:49.530797 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/registry-server/0.log" Nov 29 01:08:49 crc kubenswrapper[4931]: I1129 01:08:49.647413 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/extract-utilities/0.log" Nov 29 01:08:49 crc kubenswrapper[4931]: I1129 01:08:49.648464 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/extract-content/0.log" Nov 29 01:08:49 crc kubenswrapper[4931]: I1129 01:08:49.676328 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/extract-content/0.log" Nov 29 01:08:49 crc kubenswrapper[4931]: I1129 01:08:49.826043 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/extract-utilities/0.log" Nov 29 01:08:49 crc kubenswrapper[4931]: I1129 01:08:49.831351 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/extract-content/0.log" Nov 29 01:08:50 crc kubenswrapper[4931]: I1129 01:08:50.299221 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/registry-server/0.log" Nov 29 01:08:56 crc kubenswrapper[4931]: I1129 01:08:56.212848 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:08:57 crc kubenswrapper[4931]: I1129 01:08:57.330966 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"095e15e06fb756ce4c0646d5a8622ecc181be42bd4003d1af6ab96ca0122a6bd"} Nov 29 01:09:18 crc kubenswrapper[4931]: E1129 01:09:18.259923 4931 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.18:51880->38.102.83.18:34265: read tcp 38.102.83.18:51880->38.102.83.18:34265: read: connection reset by peer Nov 29 01:09:18 crc kubenswrapper[4931]: E1129 01:09:18.260501 4931 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.18:51880->38.102.83.18:34265: write tcp 38.102.83.18:51880->38.102.83.18:34265: write: broken pipe Nov 29 01:09:19 crc kubenswrapper[4931]: E1129 01:09:19.901156 4931 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.18:52012->38.102.83.18:34265: write tcp 38.102.83.18:52012->38.102.83.18:34265: write: broken pipe Nov 29 01:10:29 crc kubenswrapper[4931]: I1129 01:10:29.369404 4931 generic.go:334] "Generic (PLEG): container finished" podID="12d5edce-1a2f-4cce-89a4-99ed66c0cc15" containerID="6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336" exitCode=0 Nov 29 01:10:29 crc kubenswrapper[4931]: I1129 01:10:29.369527 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wn64l/must-gather-lht9f" event={"ID":"12d5edce-1a2f-4cce-89a4-99ed66c0cc15","Type":"ContainerDied","Data":"6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336"} Nov 29 01:10:29 crc kubenswrapper[4931]: I1129 01:10:29.373049 4931 scope.go:117] "RemoveContainer" containerID="6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336" Nov 29 01:10:30 crc kubenswrapper[4931]: I1129 01:10:30.173476 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wn64l_must-gather-lht9f_12d5edce-1a2f-4cce-89a4-99ed66c0cc15/gather/0.log" Nov 29 01:10:37 crc kubenswrapper[4931]: I1129 01:10:37.770875 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wn64l/must-gather-lht9f"] Nov 29 01:10:37 crc kubenswrapper[4931]: I1129 01:10:37.771713 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wn64l/must-gather-lht9f" podUID="12d5edce-1a2f-4cce-89a4-99ed66c0cc15" containerName="copy" containerID="cri-o://e0021ea63156d65ae84861c442c97a45bc1771a869423dd4cb8b53ca9edd0ced" gracePeriod=2 Nov 29 01:10:37 crc kubenswrapper[4931]: I1129 01:10:37.779706 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wn64l/must-gather-lht9f"] Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.269482 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wn64l_must-gather-lht9f_12d5edce-1a2f-4cce-89a4-99ed66c0cc15/copy/0.log" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.270609 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/must-gather-lht9f" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.386441 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/12d5edce-1a2f-4cce-89a4-99ed66c0cc15-must-gather-output\") pod \"12d5edce-1a2f-4cce-89a4-99ed66c0cc15\" (UID: \"12d5edce-1a2f-4cce-89a4-99ed66c0cc15\") " Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.386535 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqwk6\" (UniqueName: \"kubernetes.io/projected/12d5edce-1a2f-4cce-89a4-99ed66c0cc15-kube-api-access-qqwk6\") pod \"12d5edce-1a2f-4cce-89a4-99ed66c0cc15\" (UID: \"12d5edce-1a2f-4cce-89a4-99ed66c0cc15\") " Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.392607 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12d5edce-1a2f-4cce-89a4-99ed66c0cc15-kube-api-access-qqwk6" (OuterVolumeSpecName: "kube-api-access-qqwk6") pod "12d5edce-1a2f-4cce-89a4-99ed66c0cc15" (UID: "12d5edce-1a2f-4cce-89a4-99ed66c0cc15"). InnerVolumeSpecName "kube-api-access-qqwk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.479988 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wn64l_must-gather-lht9f_12d5edce-1a2f-4cce-89a4-99ed66c0cc15/copy/0.log" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.480329 4931 generic.go:334] "Generic (PLEG): container finished" podID="12d5edce-1a2f-4cce-89a4-99ed66c0cc15" containerID="e0021ea63156d65ae84861c442c97a45bc1771a869423dd4cb8b53ca9edd0ced" exitCode=143 Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.480373 4931 scope.go:117] "RemoveContainer" containerID="e0021ea63156d65ae84861c442c97a45bc1771a869423dd4cb8b53ca9edd0ced" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.480388 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wn64l/must-gather-lht9f" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.489455 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqwk6\" (UniqueName: \"kubernetes.io/projected/12d5edce-1a2f-4cce-89a4-99ed66c0cc15-kube-api-access-qqwk6\") on node \"crc\" DevicePath \"\"" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.513124 4931 scope.go:117] "RemoveContainer" containerID="6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.588824 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12d5edce-1a2f-4cce-89a4-99ed66c0cc15-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "12d5edce-1a2f-4cce-89a4-99ed66c0cc15" (UID: "12d5edce-1a2f-4cce-89a4-99ed66c0cc15"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.591262 4931 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/12d5edce-1a2f-4cce-89a4-99ed66c0cc15-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.603029 4931 scope.go:117] "RemoveContainer" containerID="e0021ea63156d65ae84861c442c97a45bc1771a869423dd4cb8b53ca9edd0ced" Nov 29 01:10:38 crc kubenswrapper[4931]: E1129 01:10:38.606664 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0021ea63156d65ae84861c442c97a45bc1771a869423dd4cb8b53ca9edd0ced\": container with ID starting with e0021ea63156d65ae84861c442c97a45bc1771a869423dd4cb8b53ca9edd0ced not found: ID does not exist" containerID="e0021ea63156d65ae84861c442c97a45bc1771a869423dd4cb8b53ca9edd0ced" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.606715 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0021ea63156d65ae84861c442c97a45bc1771a869423dd4cb8b53ca9edd0ced"} err="failed to get container status \"e0021ea63156d65ae84861c442c97a45bc1771a869423dd4cb8b53ca9edd0ced\": rpc error: code = NotFound desc = could not find container \"e0021ea63156d65ae84861c442c97a45bc1771a869423dd4cb8b53ca9edd0ced\": container with ID starting with e0021ea63156d65ae84861c442c97a45bc1771a869423dd4cb8b53ca9edd0ced not found: ID does not exist" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.606755 4931 scope.go:117] "RemoveContainer" containerID="6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336" Nov 29 01:10:38 crc kubenswrapper[4931]: E1129 01:10:38.608350 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336\": container with ID starting with 6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336 not found: ID does not exist" containerID="6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336" Nov 29 01:10:38 crc kubenswrapper[4931]: I1129 01:10:38.608388 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336"} err="failed to get container status \"6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336\": rpc error: code = NotFound desc = could not find container \"6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336\": container with ID starting with 6edd921b79d9c2473a9cd502b1922c608a60af41f0ffc14d4b41da62c9927336 not found: ID does not exist" Nov 29 01:10:39 crc kubenswrapper[4931]: I1129 01:10:39.226161 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12d5edce-1a2f-4cce-89a4-99ed66c0cc15" path="/var/lib/kubelet/pods/12d5edce-1a2f-4cce-89a4-99ed66c0cc15/volumes" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.362579 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gcgfx"] Nov 29 01:11:07 crc kubenswrapper[4931]: E1129 01:11:07.363345 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d5edce-1a2f-4cce-89a4-99ed66c0cc15" containerName="gather" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.363357 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d5edce-1a2f-4cce-89a4-99ed66c0cc15" containerName="gather" Nov 29 01:11:07 crc kubenswrapper[4931]: E1129 01:11:07.363368 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d5edce-1a2f-4cce-89a4-99ed66c0cc15" containerName="copy" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.363373 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d5edce-1a2f-4cce-89a4-99ed66c0cc15" containerName="copy" Nov 29 01:11:07 crc kubenswrapper[4931]: E1129 01:11:07.363394 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f58c30-f7fe-4141-ab70-122e133cccdc" containerName="registry-server" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.363400 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f58c30-f7fe-4141-ab70-122e133cccdc" containerName="registry-server" Nov 29 01:11:07 crc kubenswrapper[4931]: E1129 01:11:07.363432 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f58c30-f7fe-4141-ab70-122e133cccdc" containerName="extract-utilities" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.363438 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f58c30-f7fe-4141-ab70-122e133cccdc" containerName="extract-utilities" Nov 29 01:11:07 crc kubenswrapper[4931]: E1129 01:11:07.363445 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f58c30-f7fe-4141-ab70-122e133cccdc" containerName="extract-content" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.363452 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f58c30-f7fe-4141-ab70-122e133cccdc" containerName="extract-content" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.363600 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="12d5edce-1a2f-4cce-89a4-99ed66c0cc15" containerName="copy" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.363619 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="12d5edce-1a2f-4cce-89a4-99ed66c0cc15" containerName="gather" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.363629 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="59f58c30-f7fe-4141-ab70-122e133cccdc" containerName="registry-server" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.364947 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.386451 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gcgfx"] Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.520962 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g25dp\" (UniqueName: \"kubernetes.io/projected/531e6518-ff29-45b8-9321-29abd1cf6ec7-kube-api-access-g25dp\") pod \"community-operators-gcgfx\" (UID: \"531e6518-ff29-45b8-9321-29abd1cf6ec7\") " pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.521002 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531e6518-ff29-45b8-9321-29abd1cf6ec7-utilities\") pod \"community-operators-gcgfx\" (UID: \"531e6518-ff29-45b8-9321-29abd1cf6ec7\") " pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.521087 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531e6518-ff29-45b8-9321-29abd1cf6ec7-catalog-content\") pod \"community-operators-gcgfx\" (UID: \"531e6518-ff29-45b8-9321-29abd1cf6ec7\") " pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.623307 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531e6518-ff29-45b8-9321-29abd1cf6ec7-utilities\") pod \"community-operators-gcgfx\" (UID: \"531e6518-ff29-45b8-9321-29abd1cf6ec7\") " pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.623738 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g25dp\" (UniqueName: \"kubernetes.io/projected/531e6518-ff29-45b8-9321-29abd1cf6ec7-kube-api-access-g25dp\") pod \"community-operators-gcgfx\" (UID: \"531e6518-ff29-45b8-9321-29abd1cf6ec7\") " pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.623893 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531e6518-ff29-45b8-9321-29abd1cf6ec7-catalog-content\") pod \"community-operators-gcgfx\" (UID: \"531e6518-ff29-45b8-9321-29abd1cf6ec7\") " pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.624355 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/531e6518-ff29-45b8-9321-29abd1cf6ec7-utilities\") pod \"community-operators-gcgfx\" (UID: \"531e6518-ff29-45b8-9321-29abd1cf6ec7\") " pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.624584 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/531e6518-ff29-45b8-9321-29abd1cf6ec7-catalog-content\") pod \"community-operators-gcgfx\" (UID: \"531e6518-ff29-45b8-9321-29abd1cf6ec7\") " pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.653152 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g25dp\" (UniqueName: \"kubernetes.io/projected/531e6518-ff29-45b8-9321-29abd1cf6ec7-kube-api-access-g25dp\") pod \"community-operators-gcgfx\" (UID: \"531e6518-ff29-45b8-9321-29abd1cf6ec7\") " pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:07 crc kubenswrapper[4931]: I1129 01:11:07.691897 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:08 crc kubenswrapper[4931]: I1129 01:11:08.623227 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gcgfx"] Nov 29 01:11:08 crc kubenswrapper[4931]: I1129 01:11:08.885211 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gcgfx" event={"ID":"531e6518-ff29-45b8-9321-29abd1cf6ec7","Type":"ContainerStarted","Data":"e6f91be07a6ee28671a658d7b3c4b199c132abb18c87f8c18e81cf0ed30dda43"} Nov 29 01:11:08 crc kubenswrapper[4931]: I1129 01:11:08.885274 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gcgfx" event={"ID":"531e6518-ff29-45b8-9321-29abd1cf6ec7","Type":"ContainerStarted","Data":"109f845d4d20b4f4e9697afc1cc7259f1917ea73a60b93838ec96671d004bb1d"} Nov 29 01:11:08 crc kubenswrapper[4931]: I1129 01:11:08.889103 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 01:11:09 crc kubenswrapper[4931]: I1129 01:11:09.901978 4931 generic.go:334] "Generic (PLEG): container finished" podID="531e6518-ff29-45b8-9321-29abd1cf6ec7" containerID="e6f91be07a6ee28671a658d7b3c4b199c132abb18c87f8c18e81cf0ed30dda43" exitCode=0 Nov 29 01:11:09 crc kubenswrapper[4931]: I1129 01:11:09.902110 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gcgfx" event={"ID":"531e6518-ff29-45b8-9321-29abd1cf6ec7","Type":"ContainerDied","Data":"e6f91be07a6ee28671a658d7b3c4b199c132abb18c87f8c18e81cf0ed30dda43"} Nov 29 01:11:13 crc kubenswrapper[4931]: I1129 01:11:13.962140 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gcgfx" event={"ID":"531e6518-ff29-45b8-9321-29abd1cf6ec7","Type":"ContainerStarted","Data":"11efa0f7ee4475f2bb8e5674daad99e3aa3914713a34dc96a019d8041eab787f"} Nov 29 01:11:14 crc kubenswrapper[4931]: I1129 01:11:14.977501 4931 generic.go:334] "Generic (PLEG): container finished" podID="531e6518-ff29-45b8-9321-29abd1cf6ec7" containerID="11efa0f7ee4475f2bb8e5674daad99e3aa3914713a34dc96a019d8041eab787f" exitCode=0 Nov 29 01:11:14 crc kubenswrapper[4931]: I1129 01:11:14.977613 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gcgfx" event={"ID":"531e6518-ff29-45b8-9321-29abd1cf6ec7","Type":"ContainerDied","Data":"11efa0f7ee4475f2bb8e5674daad99e3aa3914713a34dc96a019d8041eab787f"} Nov 29 01:11:15 crc kubenswrapper[4931]: I1129 01:11:15.989244 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gcgfx" event={"ID":"531e6518-ff29-45b8-9321-29abd1cf6ec7","Type":"ContainerStarted","Data":"cc6ea4648a8b02652be92498554099c077ce25d34d82f8f60307cbae687c3db9"} Nov 29 01:11:16 crc kubenswrapper[4931]: I1129 01:11:16.013111 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gcgfx" podStartSLOduration=2.459861892 podStartE2EDuration="9.013091173s" podCreationTimestamp="2025-11-29 01:11:07 +0000 UTC" firstStartedPulling="2025-11-29 01:11:08.888498708 +0000 UTC m=+4282.050391970" lastFinishedPulling="2025-11-29 01:11:15.441728009 +0000 UTC m=+4288.603621251" observedRunningTime="2025-11-29 01:11:16.008927375 +0000 UTC m=+4289.170820627" watchObservedRunningTime="2025-11-29 01:11:16.013091173 +0000 UTC m=+4289.174984415" Nov 29 01:11:17 crc kubenswrapper[4931]: I1129 01:11:17.692091 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:17 crc kubenswrapper[4931]: I1129 01:11:17.692369 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:18 crc kubenswrapper[4931]: I1129 01:11:18.764861 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-gcgfx" podUID="531e6518-ff29-45b8-9321-29abd1cf6ec7" containerName="registry-server" probeResult="failure" output=< Nov 29 01:11:18 crc kubenswrapper[4931]: timeout: failed to connect service ":50051" within 1s Nov 29 01:11:18 crc kubenswrapper[4931]: > Nov 29 01:11:23 crc kubenswrapper[4931]: I1129 01:11:23.735391 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 01:11:23 crc kubenswrapper[4931]: I1129 01:11:23.736098 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 01:11:27 crc kubenswrapper[4931]: I1129 01:11:27.750494 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:27 crc kubenswrapper[4931]: I1129 01:11:27.812888 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gcgfx" Nov 29 01:11:27 crc kubenswrapper[4931]: I1129 01:11:27.904856 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gcgfx"] Nov 29 01:11:27 crc kubenswrapper[4931]: I1129 01:11:27.974280 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zn6ld"] Nov 29 01:11:27 crc kubenswrapper[4931]: I1129 01:11:27.974546 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zn6ld" podUID="c92e6619-370d-46ea-8925-a35dff600120" containerName="registry-server" containerID="cri-o://1956a442515c9e6359d595ede6dd764b77ceba18bc1a0831e6a1e08e89a241a7" gracePeriod=2 Nov 29 01:11:28 crc kubenswrapper[4931]: I1129 01:11:28.116663 4931 generic.go:334] "Generic (PLEG): container finished" podID="c92e6619-370d-46ea-8925-a35dff600120" containerID="1956a442515c9e6359d595ede6dd764b77ceba18bc1a0831e6a1e08e89a241a7" exitCode=0 Nov 29 01:11:28 crc kubenswrapper[4931]: I1129 01:11:28.116684 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zn6ld" event={"ID":"c92e6619-370d-46ea-8925-a35dff600120","Type":"ContainerDied","Data":"1956a442515c9e6359d595ede6dd764b77ceba18bc1a0831e6a1e08e89a241a7"} Nov 29 01:11:28 crc kubenswrapper[4931]: I1129 01:11:28.410213 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zn6ld" Nov 29 01:11:28 crc kubenswrapper[4931]: I1129 01:11:28.432144 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnjhp\" (UniqueName: \"kubernetes.io/projected/c92e6619-370d-46ea-8925-a35dff600120-kube-api-access-hnjhp\") pod \"c92e6619-370d-46ea-8925-a35dff600120\" (UID: \"c92e6619-370d-46ea-8925-a35dff600120\") " Nov 29 01:11:28 crc kubenswrapper[4931]: I1129 01:11:28.432281 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92e6619-370d-46ea-8925-a35dff600120-catalog-content\") pod \"c92e6619-370d-46ea-8925-a35dff600120\" (UID: \"c92e6619-370d-46ea-8925-a35dff600120\") " Nov 29 01:11:28 crc kubenswrapper[4931]: I1129 01:11:28.432542 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92e6619-370d-46ea-8925-a35dff600120-utilities\") pod \"c92e6619-370d-46ea-8925-a35dff600120\" (UID: \"c92e6619-370d-46ea-8925-a35dff600120\") " Nov 29 01:11:28 crc kubenswrapper[4931]: I1129 01:11:28.433588 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c92e6619-370d-46ea-8925-a35dff600120-utilities" (OuterVolumeSpecName: "utilities") pod "c92e6619-370d-46ea-8925-a35dff600120" (UID: "c92e6619-370d-46ea-8925-a35dff600120"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:11:28 crc kubenswrapper[4931]: I1129 01:11:28.453250 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c92e6619-370d-46ea-8925-a35dff600120-kube-api-access-hnjhp" (OuterVolumeSpecName: "kube-api-access-hnjhp") pod "c92e6619-370d-46ea-8925-a35dff600120" (UID: "c92e6619-370d-46ea-8925-a35dff600120"). InnerVolumeSpecName "kube-api-access-hnjhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:11:28 crc kubenswrapper[4931]: I1129 01:11:28.513235 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c92e6619-370d-46ea-8925-a35dff600120-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c92e6619-370d-46ea-8925-a35dff600120" (UID: "c92e6619-370d-46ea-8925-a35dff600120"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:11:28 crc kubenswrapper[4931]: I1129 01:11:28.535097 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnjhp\" (UniqueName: \"kubernetes.io/projected/c92e6619-370d-46ea-8925-a35dff600120-kube-api-access-hnjhp\") on node \"crc\" DevicePath \"\"" Nov 29 01:11:28 crc kubenswrapper[4931]: I1129 01:11:28.535132 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92e6619-370d-46ea-8925-a35dff600120-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 01:11:28 crc kubenswrapper[4931]: I1129 01:11:28.535142 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92e6619-370d-46ea-8925-a35dff600120-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 01:11:29 crc kubenswrapper[4931]: I1129 01:11:29.127332 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zn6ld" Nov 29 01:11:29 crc kubenswrapper[4931]: I1129 01:11:29.127382 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zn6ld" event={"ID":"c92e6619-370d-46ea-8925-a35dff600120","Type":"ContainerDied","Data":"1be0022430fb53125cce4ab6855e753476ccebd1577f2b02013abbc3e085f2a6"} Nov 29 01:11:29 crc kubenswrapper[4931]: I1129 01:11:29.127417 4931 scope.go:117] "RemoveContainer" containerID="1956a442515c9e6359d595ede6dd764b77ceba18bc1a0831e6a1e08e89a241a7" Nov 29 01:11:29 crc kubenswrapper[4931]: I1129 01:11:29.156160 4931 scope.go:117] "RemoveContainer" containerID="0549faedaf7419967754addc420539737620876e84f6bbe303626260f1f73594" Nov 29 01:11:29 crc kubenswrapper[4931]: I1129 01:11:29.162108 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zn6ld"] Nov 29 01:11:29 crc kubenswrapper[4931]: I1129 01:11:29.174490 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zn6ld"] Nov 29 01:11:29 crc kubenswrapper[4931]: I1129 01:11:29.204345 4931 scope.go:117] "RemoveContainer" containerID="d16769b76161e3fcb7bced18f71db25b03fdcbd9d2f4da5bafb8f6ff00931ea0" Nov 29 01:11:29 crc kubenswrapper[4931]: I1129 01:11:29.226721 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c92e6619-370d-46ea-8925-a35dff600120" path="/var/lib/kubelet/pods/c92e6619-370d-46ea-8925-a35dff600120/volumes" Nov 29 01:11:53 crc kubenswrapper[4931]: I1129 01:11:53.735087 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 01:11:53 crc kubenswrapper[4931]: I1129 01:11:53.735620 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 01:12:17 crc kubenswrapper[4931]: I1129 01:12:17.686019 4931 scope.go:117] "RemoveContainer" containerID="0472fcbea5dcb603175d817c890e59959b3eb3d1b2f1c54321630e9e5e0800dd" Nov 29 01:12:17 crc kubenswrapper[4931]: I1129 01:12:17.730582 4931 scope.go:117] "RemoveContainer" containerID="8442c11661395f41812caac6245b65f96cfb41e410279df40f9e124140fe848a" Nov 29 01:12:23 crc kubenswrapper[4931]: I1129 01:12:23.735414 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 01:12:23 crc kubenswrapper[4931]: I1129 01:12:23.736278 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 01:12:23 crc kubenswrapper[4931]: I1129 01:12:23.736351 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 01:12:23 crc kubenswrapper[4931]: I1129 01:12:23.737589 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"095e15e06fb756ce4c0646d5a8622ecc181be42bd4003d1af6ab96ca0122a6bd"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 01:12:23 crc kubenswrapper[4931]: I1129 01:12:23.737722 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://095e15e06fb756ce4c0646d5a8622ecc181be42bd4003d1af6ab96ca0122a6bd" gracePeriod=600 Nov 29 01:12:24 crc kubenswrapper[4931]: I1129 01:12:24.782601 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="095e15e06fb756ce4c0646d5a8622ecc181be42bd4003d1af6ab96ca0122a6bd" exitCode=0 Nov 29 01:12:24 crc kubenswrapper[4931]: I1129 01:12:24.782690 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"095e15e06fb756ce4c0646d5a8622ecc181be42bd4003d1af6ab96ca0122a6bd"} Nov 29 01:12:24 crc kubenswrapper[4931]: I1129 01:12:24.783196 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a"} Nov 29 01:12:24 crc kubenswrapper[4931]: I1129 01:12:24.783239 4931 scope.go:117] "RemoveContainer" containerID="29db94b844f9c63ab88b518d70777688cf9e164db42ccb76732b4341e5ddf094" Nov 29 01:13:25 crc kubenswrapper[4931]: I1129 01:13:25.857732 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zn82b/must-gather-gtd7t"] Nov 29 01:13:25 crc kubenswrapper[4931]: E1129 01:13:25.858634 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92e6619-370d-46ea-8925-a35dff600120" containerName="extract-content" Nov 29 01:13:25 crc kubenswrapper[4931]: I1129 01:13:25.858647 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92e6619-370d-46ea-8925-a35dff600120" containerName="extract-content" Nov 29 01:13:25 crc kubenswrapper[4931]: E1129 01:13:25.858663 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92e6619-370d-46ea-8925-a35dff600120" containerName="registry-server" Nov 29 01:13:25 crc kubenswrapper[4931]: I1129 01:13:25.858668 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92e6619-370d-46ea-8925-a35dff600120" containerName="registry-server" Nov 29 01:13:25 crc kubenswrapper[4931]: E1129 01:13:25.858683 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92e6619-370d-46ea-8925-a35dff600120" containerName="extract-utilities" Nov 29 01:13:25 crc kubenswrapper[4931]: I1129 01:13:25.858690 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92e6619-370d-46ea-8925-a35dff600120" containerName="extract-utilities" Nov 29 01:13:25 crc kubenswrapper[4931]: I1129 01:13:25.858908 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="c92e6619-370d-46ea-8925-a35dff600120" containerName="registry-server" Nov 29 01:13:25 crc kubenswrapper[4931]: I1129 01:13:25.860985 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/must-gather-gtd7t" Nov 29 01:13:25 crc kubenswrapper[4931]: I1129 01:13:25.865942 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zn82b"/"kube-root-ca.crt" Nov 29 01:13:25 crc kubenswrapper[4931]: I1129 01:13:25.866240 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zn82b"/"openshift-service-ca.crt" Nov 29 01:13:25 crc kubenswrapper[4931]: I1129 01:13:25.876000 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zn82b/must-gather-gtd7t"] Nov 29 01:13:25 crc kubenswrapper[4931]: I1129 01:13:25.952110 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn828\" (UniqueName: \"kubernetes.io/projected/143aa3c5-6047-4d24-affe-698ec54aae1b-kube-api-access-pn828\") pod \"must-gather-gtd7t\" (UID: \"143aa3c5-6047-4d24-affe-698ec54aae1b\") " pod="openshift-must-gather-zn82b/must-gather-gtd7t" Nov 29 01:13:25 crc kubenswrapper[4931]: I1129 01:13:25.952181 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/143aa3c5-6047-4d24-affe-698ec54aae1b-must-gather-output\") pod \"must-gather-gtd7t\" (UID: \"143aa3c5-6047-4d24-affe-698ec54aae1b\") " pod="openshift-must-gather-zn82b/must-gather-gtd7t" Nov 29 01:13:26 crc kubenswrapper[4931]: I1129 01:13:26.053999 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn828\" (UniqueName: \"kubernetes.io/projected/143aa3c5-6047-4d24-affe-698ec54aae1b-kube-api-access-pn828\") pod \"must-gather-gtd7t\" (UID: \"143aa3c5-6047-4d24-affe-698ec54aae1b\") " pod="openshift-must-gather-zn82b/must-gather-gtd7t" Nov 29 01:13:26 crc kubenswrapper[4931]: I1129 01:13:26.054073 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/143aa3c5-6047-4d24-affe-698ec54aae1b-must-gather-output\") pod \"must-gather-gtd7t\" (UID: \"143aa3c5-6047-4d24-affe-698ec54aae1b\") " pod="openshift-must-gather-zn82b/must-gather-gtd7t" Nov 29 01:13:26 crc kubenswrapper[4931]: I1129 01:13:26.054772 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/143aa3c5-6047-4d24-affe-698ec54aae1b-must-gather-output\") pod \"must-gather-gtd7t\" (UID: \"143aa3c5-6047-4d24-affe-698ec54aae1b\") " pod="openshift-must-gather-zn82b/must-gather-gtd7t" Nov 29 01:13:26 crc kubenswrapper[4931]: I1129 01:13:26.072576 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn828\" (UniqueName: \"kubernetes.io/projected/143aa3c5-6047-4d24-affe-698ec54aae1b-kube-api-access-pn828\") pod \"must-gather-gtd7t\" (UID: \"143aa3c5-6047-4d24-affe-698ec54aae1b\") " pod="openshift-must-gather-zn82b/must-gather-gtd7t" Nov 29 01:13:26 crc kubenswrapper[4931]: I1129 01:13:26.201801 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/must-gather-gtd7t" Nov 29 01:13:26 crc kubenswrapper[4931]: I1129 01:13:26.664985 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zn82b/must-gather-gtd7t"] Nov 29 01:13:27 crc kubenswrapper[4931]: I1129 01:13:27.477200 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn82b/must-gather-gtd7t" event={"ID":"143aa3c5-6047-4d24-affe-698ec54aae1b","Type":"ContainerStarted","Data":"f15c4ff22dfd93864b6f6873621985d5fbcdfa3122c761eaef76fc133ba36ae3"} Nov 29 01:13:27 crc kubenswrapper[4931]: I1129 01:13:27.477466 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn82b/must-gather-gtd7t" event={"ID":"143aa3c5-6047-4d24-affe-698ec54aae1b","Type":"ContainerStarted","Data":"1e7c7462e436cbe72c15a1a51e6f22a236f8a8fda49988622ad7a7b0feaf5294"} Nov 29 01:13:28 crc kubenswrapper[4931]: I1129 01:13:28.491995 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn82b/must-gather-gtd7t" event={"ID":"143aa3c5-6047-4d24-affe-698ec54aae1b","Type":"ContainerStarted","Data":"7d2931064538572450525d2e1fe19643cdb225a7ea2579dcf8a0009599f85af8"} Nov 29 01:13:30 crc kubenswrapper[4931]: E1129 01:13:30.526985 4931 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.18:37088->38.102.83.18:34265: write tcp 38.102.83.18:37088->38.102.83.18:34265: write: broken pipe Nov 29 01:13:31 crc kubenswrapper[4931]: I1129 01:13:31.228323 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zn82b/must-gather-gtd7t" podStartSLOduration=6.228296237 podStartE2EDuration="6.228296237s" podCreationTimestamp="2025-11-29 01:13:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 01:13:28.513295541 +0000 UTC m=+4421.675188793" watchObservedRunningTime="2025-11-29 01:13:31.228296237 +0000 UTC m=+4424.390189469" Nov 29 01:13:31 crc kubenswrapper[4931]: I1129 01:13:31.234046 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zn82b/crc-debug-c2q8b"] Nov 29 01:13:31 crc kubenswrapper[4931]: I1129 01:13:31.235283 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/crc-debug-c2q8b" Nov 29 01:13:31 crc kubenswrapper[4931]: I1129 01:13:31.237344 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-zn82b"/"default-dockercfg-hpmvs" Nov 29 01:13:31 crc kubenswrapper[4931]: I1129 01:13:31.368565 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc-host\") pod \"crc-debug-c2q8b\" (UID: \"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc\") " pod="openshift-must-gather-zn82b/crc-debug-c2q8b" Nov 29 01:13:31 crc kubenswrapper[4931]: I1129 01:13:31.368949 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbkn7\" (UniqueName: \"kubernetes.io/projected/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc-kube-api-access-sbkn7\") pod \"crc-debug-c2q8b\" (UID: \"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc\") " pod="openshift-must-gather-zn82b/crc-debug-c2q8b" Nov 29 01:13:31 crc kubenswrapper[4931]: I1129 01:13:31.470595 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc-host\") pod \"crc-debug-c2q8b\" (UID: \"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc\") " pod="openshift-must-gather-zn82b/crc-debug-c2q8b" Nov 29 01:13:31 crc kubenswrapper[4931]: I1129 01:13:31.470662 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbkn7\" (UniqueName: \"kubernetes.io/projected/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc-kube-api-access-sbkn7\") pod \"crc-debug-c2q8b\" (UID: \"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc\") " pod="openshift-must-gather-zn82b/crc-debug-c2q8b" Nov 29 01:13:31 crc kubenswrapper[4931]: I1129 01:13:31.470791 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc-host\") pod \"crc-debug-c2q8b\" (UID: \"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc\") " pod="openshift-must-gather-zn82b/crc-debug-c2q8b" Nov 29 01:13:31 crc kubenswrapper[4931]: I1129 01:13:31.496850 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbkn7\" (UniqueName: \"kubernetes.io/projected/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc-kube-api-access-sbkn7\") pod \"crc-debug-c2q8b\" (UID: \"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc\") " pod="openshift-must-gather-zn82b/crc-debug-c2q8b" Nov 29 01:13:31 crc kubenswrapper[4931]: I1129 01:13:31.556476 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/crc-debug-c2q8b" Nov 29 01:13:31 crc kubenswrapper[4931]: W1129 01:13:31.593740 4931 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cd78bc3_11a9_4bbe_923f_738cdd6db7bc.slice/crio-7d354d98973ca59d9c207d34c65865375aec45b80667a7e14db40b154a60b7b8 WatchSource:0}: Error finding container 7d354d98973ca59d9c207d34c65865375aec45b80667a7e14db40b154a60b7b8: Status 404 returned error can't find the container with id 7d354d98973ca59d9c207d34c65865375aec45b80667a7e14db40b154a60b7b8 Nov 29 01:13:32 crc kubenswrapper[4931]: I1129 01:13:32.529086 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn82b/crc-debug-c2q8b" event={"ID":"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc","Type":"ContainerStarted","Data":"8479715e2efd53afe5240c55505ef9febf3efd4a84ed27fedf6cf502f6272466"} Nov 29 01:13:32 crc kubenswrapper[4931]: I1129 01:13:32.529672 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn82b/crc-debug-c2q8b" event={"ID":"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc","Type":"ContainerStarted","Data":"7d354d98973ca59d9c207d34c65865375aec45b80667a7e14db40b154a60b7b8"} Nov 29 01:13:32 crc kubenswrapper[4931]: I1129 01:13:32.550692 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zn82b/crc-debug-c2q8b" podStartSLOduration=1.550666841 podStartE2EDuration="1.550666841s" podCreationTimestamp="2025-11-29 01:13:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 01:13:32.541679958 +0000 UTC m=+4425.703573230" watchObservedRunningTime="2025-11-29 01:13:32.550666841 +0000 UTC m=+4425.712560073" Nov 29 01:14:05 crc kubenswrapper[4931]: I1129 01:14:05.848624 4931 generic.go:334] "Generic (PLEG): container finished" podID="0cd78bc3-11a9-4bbe-923f-738cdd6db7bc" containerID="8479715e2efd53afe5240c55505ef9febf3efd4a84ed27fedf6cf502f6272466" exitCode=0 Nov 29 01:14:05 crc kubenswrapper[4931]: I1129 01:14:05.849144 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn82b/crc-debug-c2q8b" event={"ID":"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc","Type":"ContainerDied","Data":"8479715e2efd53afe5240c55505ef9febf3efd4a84ed27fedf6cf502f6272466"} Nov 29 01:14:06 crc kubenswrapper[4931]: I1129 01:14:06.976629 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/crc-debug-c2q8b" Nov 29 01:14:07 crc kubenswrapper[4931]: I1129 01:14:07.005940 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zn82b/crc-debug-c2q8b"] Nov 29 01:14:07 crc kubenswrapper[4931]: I1129 01:14:07.014431 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zn82b/crc-debug-c2q8b"] Nov 29 01:14:07 crc kubenswrapper[4931]: I1129 01:14:07.149486 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc-host\") pod \"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc\" (UID: \"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc\") " Nov 29 01:14:07 crc kubenswrapper[4931]: I1129 01:14:07.149647 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc-host" (OuterVolumeSpecName: "host") pod "0cd78bc3-11a9-4bbe-923f-738cdd6db7bc" (UID: "0cd78bc3-11a9-4bbe-923f-738cdd6db7bc"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 01:14:07 crc kubenswrapper[4931]: I1129 01:14:07.149883 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbkn7\" (UniqueName: \"kubernetes.io/projected/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc-kube-api-access-sbkn7\") pod \"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc\" (UID: \"0cd78bc3-11a9-4bbe-923f-738cdd6db7bc\") " Nov 29 01:14:07 crc kubenswrapper[4931]: I1129 01:14:07.150356 4931 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc-host\") on node \"crc\" DevicePath \"\"" Nov 29 01:14:07 crc kubenswrapper[4931]: I1129 01:14:07.157232 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc-kube-api-access-sbkn7" (OuterVolumeSpecName: "kube-api-access-sbkn7") pod "0cd78bc3-11a9-4bbe-923f-738cdd6db7bc" (UID: "0cd78bc3-11a9-4bbe-923f-738cdd6db7bc"). InnerVolumeSpecName "kube-api-access-sbkn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:14:07 crc kubenswrapper[4931]: I1129 01:14:07.227983 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cd78bc3-11a9-4bbe-923f-738cdd6db7bc" path="/var/lib/kubelet/pods/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc/volumes" Nov 29 01:14:07 crc kubenswrapper[4931]: I1129 01:14:07.251963 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbkn7\" (UniqueName: \"kubernetes.io/projected/0cd78bc3-11a9-4bbe-923f-738cdd6db7bc-kube-api-access-sbkn7\") on node \"crc\" DevicePath \"\"" Nov 29 01:14:07 crc kubenswrapper[4931]: I1129 01:14:07.865712 4931 scope.go:117] "RemoveContainer" containerID="8479715e2efd53afe5240c55505ef9febf3efd4a84ed27fedf6cf502f6272466" Nov 29 01:14:07 crc kubenswrapper[4931]: I1129 01:14:07.865758 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/crc-debug-c2q8b" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.194822 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zn82b/crc-debug-s6kvn"] Nov 29 01:14:08 crc kubenswrapper[4931]: E1129 01:14:08.195187 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cd78bc3-11a9-4bbe-923f-738cdd6db7bc" containerName="container-00" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.195199 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cd78bc3-11a9-4bbe-923f-738cdd6db7bc" containerName="container-00" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.195422 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cd78bc3-11a9-4bbe-923f-738cdd6db7bc" containerName="container-00" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.196039 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/crc-debug-s6kvn" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.197644 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-zn82b"/"default-dockercfg-hpmvs" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.386216 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x659s\" (UniqueName: \"kubernetes.io/projected/cc42cc67-1c85-4e1e-b0fd-957682b186c8-kube-api-access-x659s\") pod \"crc-debug-s6kvn\" (UID: \"cc42cc67-1c85-4e1e-b0fd-957682b186c8\") " pod="openshift-must-gather-zn82b/crc-debug-s6kvn" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.386549 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc42cc67-1c85-4e1e-b0fd-957682b186c8-host\") pod \"crc-debug-s6kvn\" (UID: \"cc42cc67-1c85-4e1e-b0fd-957682b186c8\") " pod="openshift-must-gather-zn82b/crc-debug-s6kvn" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.488676 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x659s\" (UniqueName: \"kubernetes.io/projected/cc42cc67-1c85-4e1e-b0fd-957682b186c8-kube-api-access-x659s\") pod \"crc-debug-s6kvn\" (UID: \"cc42cc67-1c85-4e1e-b0fd-957682b186c8\") " pod="openshift-must-gather-zn82b/crc-debug-s6kvn" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.489117 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc42cc67-1c85-4e1e-b0fd-957682b186c8-host\") pod \"crc-debug-s6kvn\" (UID: \"cc42cc67-1c85-4e1e-b0fd-957682b186c8\") " pod="openshift-must-gather-zn82b/crc-debug-s6kvn" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.489185 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc42cc67-1c85-4e1e-b0fd-957682b186c8-host\") pod \"crc-debug-s6kvn\" (UID: \"cc42cc67-1c85-4e1e-b0fd-957682b186c8\") " pod="openshift-must-gather-zn82b/crc-debug-s6kvn" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.511758 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x659s\" (UniqueName: \"kubernetes.io/projected/cc42cc67-1c85-4e1e-b0fd-957682b186c8-kube-api-access-x659s\") pod \"crc-debug-s6kvn\" (UID: \"cc42cc67-1c85-4e1e-b0fd-957682b186c8\") " pod="openshift-must-gather-zn82b/crc-debug-s6kvn" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.513666 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/crc-debug-s6kvn" Nov 29 01:14:08 crc kubenswrapper[4931]: I1129 01:14:08.878503 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn82b/crc-debug-s6kvn" event={"ID":"cc42cc67-1c85-4e1e-b0fd-957682b186c8","Type":"ContainerStarted","Data":"4160b06649248c7cec6bed788604576e2ac99c94dd9e414c9bba63c50c0afa6b"} Nov 29 01:14:11 crc kubenswrapper[4931]: I1129 01:14:11.920098 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn82b/crc-debug-s6kvn" event={"ID":"cc42cc67-1c85-4e1e-b0fd-957682b186c8","Type":"ContainerStarted","Data":"6930cc45ec83f44b152fedfbab1ba38e3eb93d4d9d938bab59f66e48da500a23"} Nov 29 01:14:11 crc kubenswrapper[4931]: I1129 01:14:11.943772 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zn82b/crc-debug-s6kvn" podStartSLOduration=3.943748714 podStartE2EDuration="3.943748714s" podCreationTimestamp="2025-11-29 01:14:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-29 01:14:11.93401888 +0000 UTC m=+4465.095912112" watchObservedRunningTime="2025-11-29 01:14:11.943748714 +0000 UTC m=+4465.105641946" Nov 29 01:14:12 crc kubenswrapper[4931]: I1129 01:14:12.953858 4931 generic.go:334] "Generic (PLEG): container finished" podID="cc42cc67-1c85-4e1e-b0fd-957682b186c8" containerID="6930cc45ec83f44b152fedfbab1ba38e3eb93d4d9d938bab59f66e48da500a23" exitCode=0 Nov 29 01:14:12 crc kubenswrapper[4931]: I1129 01:14:12.953949 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn82b/crc-debug-s6kvn" event={"ID":"cc42cc67-1c85-4e1e-b0fd-957682b186c8","Type":"ContainerDied","Data":"6930cc45ec83f44b152fedfbab1ba38e3eb93d4d9d938bab59f66e48da500a23"} Nov 29 01:14:14 crc kubenswrapper[4931]: I1129 01:14:14.091245 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/crc-debug-s6kvn" Nov 29 01:14:14 crc kubenswrapper[4931]: I1129 01:14:14.130961 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zn82b/crc-debug-s6kvn"] Nov 29 01:14:14 crc kubenswrapper[4931]: I1129 01:14:14.138053 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zn82b/crc-debug-s6kvn"] Nov 29 01:14:14 crc kubenswrapper[4931]: I1129 01:14:14.198078 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x659s\" (UniqueName: \"kubernetes.io/projected/cc42cc67-1c85-4e1e-b0fd-957682b186c8-kube-api-access-x659s\") pod \"cc42cc67-1c85-4e1e-b0fd-957682b186c8\" (UID: \"cc42cc67-1c85-4e1e-b0fd-957682b186c8\") " Nov 29 01:14:14 crc kubenswrapper[4931]: I1129 01:14:14.198225 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc42cc67-1c85-4e1e-b0fd-957682b186c8-host\") pod \"cc42cc67-1c85-4e1e-b0fd-957682b186c8\" (UID: \"cc42cc67-1c85-4e1e-b0fd-957682b186c8\") " Nov 29 01:14:14 crc kubenswrapper[4931]: I1129 01:14:14.198361 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc42cc67-1c85-4e1e-b0fd-957682b186c8-host" (OuterVolumeSpecName: "host") pod "cc42cc67-1c85-4e1e-b0fd-957682b186c8" (UID: "cc42cc67-1c85-4e1e-b0fd-957682b186c8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 01:14:14 crc kubenswrapper[4931]: I1129 01:14:14.198711 4931 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc42cc67-1c85-4e1e-b0fd-957682b186c8-host\") on node \"crc\" DevicePath \"\"" Nov 29 01:14:14 crc kubenswrapper[4931]: I1129 01:14:14.203423 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc42cc67-1c85-4e1e-b0fd-957682b186c8-kube-api-access-x659s" (OuterVolumeSpecName: "kube-api-access-x659s") pod "cc42cc67-1c85-4e1e-b0fd-957682b186c8" (UID: "cc42cc67-1c85-4e1e-b0fd-957682b186c8"). InnerVolumeSpecName "kube-api-access-x659s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:14:14 crc kubenswrapper[4931]: I1129 01:14:14.301738 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x659s\" (UniqueName: \"kubernetes.io/projected/cc42cc67-1c85-4e1e-b0fd-957682b186c8-kube-api-access-x659s\") on node \"crc\" DevicePath \"\"" Nov 29 01:14:14 crc kubenswrapper[4931]: I1129 01:14:14.985476 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4160b06649248c7cec6bed788604576e2ac99c94dd9e414c9bba63c50c0afa6b" Nov 29 01:14:14 crc kubenswrapper[4931]: I1129 01:14:14.985550 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/crc-debug-s6kvn" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.221913 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc42cc67-1c85-4e1e-b0fd-957682b186c8" path="/var/lib/kubelet/pods/cc42cc67-1c85-4e1e-b0fd-957682b186c8/volumes" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.307678 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zn82b/crc-debug-chkdp"] Nov 29 01:14:15 crc kubenswrapper[4931]: E1129 01:14:15.308137 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc42cc67-1c85-4e1e-b0fd-957682b186c8" containerName="container-00" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.308162 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc42cc67-1c85-4e1e-b0fd-957682b186c8" containerName="container-00" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.308355 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc42cc67-1c85-4e1e-b0fd-957682b186c8" containerName="container-00" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.308971 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/crc-debug-chkdp" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.311346 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-zn82b"/"default-dockercfg-hpmvs" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.420378 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0a25d298-51d7-4759-9e62-7157d0e4c584-host\") pod \"crc-debug-chkdp\" (UID: \"0a25d298-51d7-4759-9e62-7157d0e4c584\") " pod="openshift-must-gather-zn82b/crc-debug-chkdp" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.420442 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt7fl\" (UniqueName: \"kubernetes.io/projected/0a25d298-51d7-4759-9e62-7157d0e4c584-kube-api-access-lt7fl\") pod \"crc-debug-chkdp\" (UID: \"0a25d298-51d7-4759-9e62-7157d0e4c584\") " pod="openshift-must-gather-zn82b/crc-debug-chkdp" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.523018 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0a25d298-51d7-4759-9e62-7157d0e4c584-host\") pod \"crc-debug-chkdp\" (UID: \"0a25d298-51d7-4759-9e62-7157d0e4c584\") " pod="openshift-must-gather-zn82b/crc-debug-chkdp" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.523105 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt7fl\" (UniqueName: \"kubernetes.io/projected/0a25d298-51d7-4759-9e62-7157d0e4c584-kube-api-access-lt7fl\") pod \"crc-debug-chkdp\" (UID: \"0a25d298-51d7-4759-9e62-7157d0e4c584\") " pod="openshift-must-gather-zn82b/crc-debug-chkdp" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.523832 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0a25d298-51d7-4759-9e62-7157d0e4c584-host\") pod \"crc-debug-chkdp\" (UID: \"0a25d298-51d7-4759-9e62-7157d0e4c584\") " pod="openshift-must-gather-zn82b/crc-debug-chkdp" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.545623 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt7fl\" (UniqueName: \"kubernetes.io/projected/0a25d298-51d7-4759-9e62-7157d0e4c584-kube-api-access-lt7fl\") pod \"crc-debug-chkdp\" (UID: \"0a25d298-51d7-4759-9e62-7157d0e4c584\") " pod="openshift-must-gather-zn82b/crc-debug-chkdp" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.626678 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/crc-debug-chkdp" Nov 29 01:14:15 crc kubenswrapper[4931]: I1129 01:14:15.994045 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn82b/crc-debug-chkdp" event={"ID":"0a25d298-51d7-4759-9e62-7157d0e4c584","Type":"ContainerStarted","Data":"cc403064e5f30dde75eba17a4ee885a377df6ec26824d6eb47ddcd6e620a3209"} Nov 29 01:14:18 crc kubenswrapper[4931]: I1129 01:14:18.022118 4931 generic.go:334] "Generic (PLEG): container finished" podID="0a25d298-51d7-4759-9e62-7157d0e4c584" containerID="bdd8c9f5d295b94cf883570747ce41f5cfd97c1e8071ee47a3c810a31e84da97" exitCode=0 Nov 29 01:14:18 crc kubenswrapper[4931]: I1129 01:14:18.022179 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn82b/crc-debug-chkdp" event={"ID":"0a25d298-51d7-4759-9e62-7157d0e4c584","Type":"ContainerDied","Data":"bdd8c9f5d295b94cf883570747ce41f5cfd97c1e8071ee47a3c810a31e84da97"} Nov 29 01:14:19 crc kubenswrapper[4931]: I1129 01:14:19.093959 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zn82b/crc-debug-chkdp"] Nov 29 01:14:19 crc kubenswrapper[4931]: I1129 01:14:19.110736 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zn82b/crc-debug-chkdp"] Nov 29 01:14:20 crc kubenswrapper[4931]: I1129 01:14:20.494068 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/crc-debug-chkdp" Nov 29 01:14:20 crc kubenswrapper[4931]: I1129 01:14:20.636861 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0a25d298-51d7-4759-9e62-7157d0e4c584-host\") pod \"0a25d298-51d7-4759-9e62-7157d0e4c584\" (UID: \"0a25d298-51d7-4759-9e62-7157d0e4c584\") " Nov 29 01:14:20 crc kubenswrapper[4931]: I1129 01:14:20.636984 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a25d298-51d7-4759-9e62-7157d0e4c584-host" (OuterVolumeSpecName: "host") pod "0a25d298-51d7-4759-9e62-7157d0e4c584" (UID: "0a25d298-51d7-4759-9e62-7157d0e4c584"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 29 01:14:20 crc kubenswrapper[4931]: I1129 01:14:20.637168 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt7fl\" (UniqueName: \"kubernetes.io/projected/0a25d298-51d7-4759-9e62-7157d0e4c584-kube-api-access-lt7fl\") pod \"0a25d298-51d7-4759-9e62-7157d0e4c584\" (UID: \"0a25d298-51d7-4759-9e62-7157d0e4c584\") " Nov 29 01:14:20 crc kubenswrapper[4931]: I1129 01:14:20.637842 4931 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0a25d298-51d7-4759-9e62-7157d0e4c584-host\") on node \"crc\" DevicePath \"\"" Nov 29 01:14:20 crc kubenswrapper[4931]: I1129 01:14:20.643218 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a25d298-51d7-4759-9e62-7157d0e4c584-kube-api-access-lt7fl" (OuterVolumeSpecName: "kube-api-access-lt7fl") pod "0a25d298-51d7-4759-9e62-7157d0e4c584" (UID: "0a25d298-51d7-4759-9e62-7157d0e4c584"). InnerVolumeSpecName "kube-api-access-lt7fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:14:20 crc kubenswrapper[4931]: I1129 01:14:20.739966 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt7fl\" (UniqueName: \"kubernetes.io/projected/0a25d298-51d7-4759-9e62-7157d0e4c584-kube-api-access-lt7fl\") on node \"crc\" DevicePath \"\"" Nov 29 01:14:21 crc kubenswrapper[4931]: I1129 01:14:21.058977 4931 scope.go:117] "RemoveContainer" containerID="bdd8c9f5d295b94cf883570747ce41f5cfd97c1e8071ee47a3c810a31e84da97" Nov 29 01:14:21 crc kubenswrapper[4931]: I1129 01:14:21.059053 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/crc-debug-chkdp" Nov 29 01:14:21 crc kubenswrapper[4931]: I1129 01:14:21.224620 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a25d298-51d7-4759-9e62-7157d0e4c584" path="/var/lib/kubelet/pods/0a25d298-51d7-4759-9e62-7157d0e4c584/volumes" Nov 29 01:14:44 crc kubenswrapper[4931]: I1129 01:14:44.063612 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d5fd9f77d-c4gvt_d6aa6c84-7c6c-488e-bf3b-7393ad629ed4/barbican-api/0.log" Nov 29 01:14:44 crc kubenswrapper[4931]: I1129 01:14:44.165576 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d5fd9f77d-c4gvt_d6aa6c84-7c6c-488e-bf3b-7393ad629ed4/barbican-api-log/0.log" Nov 29 01:14:44 crc kubenswrapper[4931]: I1129 01:14:44.256210 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b65859cc8-zkpcj_5941c09d-aeac-449b-bd1f-b8d590ab5ecc/barbican-keystone-listener/0.log" Nov 29 01:14:44 crc kubenswrapper[4931]: I1129 01:14:44.269303 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b65859cc8-zkpcj_5941c09d-aeac-449b-bd1f-b8d590ab5ecc/barbican-keystone-listener-log/0.log" Nov 29 01:14:44 crc kubenswrapper[4931]: I1129 01:14:44.448998 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5644b469ff-dnc8z_2ba68cea-6add-4e52-96ad-458f708a02e0/barbican-worker-log/0.log" Nov 29 01:14:44 crc kubenswrapper[4931]: I1129 01:14:44.466362 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5644b469ff-dnc8z_2ba68cea-6add-4e52-96ad-458f708a02e0/barbican-worker/0.log" Nov 29 01:14:44 crc kubenswrapper[4931]: I1129 01:14:44.594394 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-zn4cv_6a0a53a8-99d7-4a99-bdf0-67f72b8cca71/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:44 crc kubenswrapper[4931]: I1129 01:14:44.680500 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ecafd71-ca29-4382-9d56-d1e7422d113b/ceilometer-central-agent/0.log" Nov 29 01:14:44 crc kubenswrapper[4931]: I1129 01:14:44.784084 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ecafd71-ca29-4382-9d56-d1e7422d113b/ceilometer-notification-agent/0.log" Nov 29 01:14:44 crc kubenswrapper[4931]: I1129 01:14:44.813098 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ecafd71-ca29-4382-9d56-d1e7422d113b/proxy-httpd/0.log" Nov 29 01:14:44 crc kubenswrapper[4931]: I1129 01:14:44.873998 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3ecafd71-ca29-4382-9d56-d1e7422d113b/sg-core/0.log" Nov 29 01:14:45 crc kubenswrapper[4931]: I1129 01:14:45.041578 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_80c8c44d-d0c0-4d7a-9063-9915b1c4aea6/cinder-api/0.log" Nov 29 01:14:45 crc kubenswrapper[4931]: I1129 01:14:45.093947 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_80c8c44d-d0c0-4d7a-9063-9915b1c4aea6/cinder-api-log/0.log" Nov 29 01:14:45 crc kubenswrapper[4931]: I1129 01:14:45.182008 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_06e9c000-a994-4023-a6a1-eaf84b161005/cinder-scheduler/0.log" Nov 29 01:14:45 crc kubenswrapper[4931]: I1129 01:14:45.262360 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_06e9c000-a994-4023-a6a1-eaf84b161005/probe/0.log" Nov 29 01:14:45 crc kubenswrapper[4931]: I1129 01:14:45.296699 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-tnqzp_3d56a6b7-f516-4675-b2bc-50372feb503d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:45 crc kubenswrapper[4931]: I1129 01:14:45.448567 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-brwbq_5da7ac90-6a60-4473-b8b8-5531d4c2ac68/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:45 crc kubenswrapper[4931]: I1129 01:14:45.541157 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-578c4b6ff9-7b25t_76e2de18-8995-446b-aaf0-427974e0334b/init/0.log" Nov 29 01:14:45 crc kubenswrapper[4931]: I1129 01:14:45.719868 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-578c4b6ff9-7b25t_76e2de18-8995-446b-aaf0-427974e0334b/init/0.log" Nov 29 01:14:45 crc kubenswrapper[4931]: I1129 01:14:45.746916 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-578c4b6ff9-7b25t_76e2de18-8995-446b-aaf0-427974e0334b/dnsmasq-dns/0.log" Nov 29 01:14:45 crc kubenswrapper[4931]: I1129 01:14:45.795206 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-k4sf6_935c6e60-45b7-4859-bbd0-b147fa600465/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:46 crc kubenswrapper[4931]: I1129 01:14:46.021768 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ea6acd51-8e93-4fdd-9535-84c7e6d2cee6/glance-httpd/0.log" Nov 29 01:14:46 crc kubenswrapper[4931]: I1129 01:14:46.041452 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ea6acd51-8e93-4fdd-9535-84c7e6d2cee6/glance-log/0.log" Nov 29 01:14:46 crc kubenswrapper[4931]: I1129 01:14:46.144527 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_78acdbbf-4fab-4641-86e8-49f124b8247e/glance-httpd/0.log" Nov 29 01:14:46 crc kubenswrapper[4931]: I1129 01:14:46.171592 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_78acdbbf-4fab-4641-86e8-49f124b8247e/glance-log/0.log" Nov 29 01:14:46 crc kubenswrapper[4931]: I1129 01:14:46.276696 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-79d99cf788-wwqzd_4003b615-a779-45a5-b95b-913c80226f4a/horizon/0.log" Nov 29 01:14:46 crc kubenswrapper[4931]: I1129 01:14:46.520833 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-v2dj7_88f61c01-c5da-4d84-a484-48f4e3b4e3e4/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:46 crc kubenswrapper[4931]: I1129 01:14:46.658190 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-kgz5j_cb09a671-9026-4aec-ba97-911979c3f0d7/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:46 crc kubenswrapper[4931]: I1129 01:14:46.716205 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-79d99cf788-wwqzd_4003b615-a779-45a5-b95b-913c80226f4a/horizon-log/0.log" Nov 29 01:14:46 crc kubenswrapper[4931]: I1129 01:14:46.855308 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-84d76c555c-shgd6_aff73224-6284-4ef1-ace3-06eb11e904b8/keystone-api/0.log" Nov 29 01:14:46 crc kubenswrapper[4931]: I1129 01:14:46.979852 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29406301-hhnfj_55460ba8-a9d1-4ef2-9bcf-e08d863698fa/keystone-cron/0.log" Nov 29 01:14:47 crc kubenswrapper[4931]: I1129 01:14:47.046468 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b3edf21b-40ad-40e1-98db-78f609a9b4bc/kube-state-metrics/0.log" Nov 29 01:14:47 crc kubenswrapper[4931]: I1129 01:14:47.181837 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-r5958_ca54c57b-1fac-46bf-aa71-830b5417fca0/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:47 crc kubenswrapper[4931]: I1129 01:14:47.559221 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bcd7d89b9-l4sjs_e1ff7fdf-846a-4772-b8d8-fee7b15eed58/neutron-api/0.log" Nov 29 01:14:47 crc kubenswrapper[4931]: I1129 01:14:47.633291 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bcd7d89b9-l4sjs_e1ff7fdf-846a-4772-b8d8-fee7b15eed58/neutron-httpd/0.log" Nov 29 01:14:47 crc kubenswrapper[4931]: I1129 01:14:47.813337 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-9hz52_054b5e53-28bc-4222-9f59-ad0ebd2cf767/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:48 crc kubenswrapper[4931]: I1129 01:14:48.318264 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_611072c6-340e-49fc-870e-8a1e56f2cc90/nova-api-log/0.log" Nov 29 01:14:48 crc kubenswrapper[4931]: I1129 01:14:48.341040 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_2dcc4f63-f88c-436a-96a6-cae43c016e0c/nova-cell0-conductor-conductor/0.log" Nov 29 01:14:48 crc kubenswrapper[4931]: I1129 01:14:48.649367 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ecb51737-365b-4866-9d46-3e3aa5729b38/nova-cell1-conductor-conductor/0.log" Nov 29 01:14:48 crc kubenswrapper[4931]: I1129 01:14:48.728624 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_82ad9af8-53d9-48fc-b251-3b580018e51d/nova-cell1-novncproxy-novncproxy/0.log" Nov 29 01:14:48 crc kubenswrapper[4931]: I1129 01:14:48.769785 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_611072c6-340e-49fc-870e-8a1e56f2cc90/nova-api-api/0.log" Nov 29 01:14:49 crc kubenswrapper[4931]: I1129 01:14:49.314301 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-srlj9_aa634449-4a1f-4827-a318-5502c395886e/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:49 crc kubenswrapper[4931]: I1129 01:14:49.373885 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_6440833b-24fe-4da5-afae-6111fc7e760e/nova-metadata-log/0.log" Nov 29 01:14:49 crc kubenswrapper[4931]: I1129 01:14:49.735110 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_f9ed9e15-1e99-44d2-9b7d-1f422e79f4c2/nova-scheduler-scheduler/0.log" Nov 29 01:14:49 crc kubenswrapper[4931]: I1129 01:14:49.849301 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2b4d256-9b86-48e3-98d8-266271a338dd/mysql-bootstrap/0.log" Nov 29 01:14:50 crc kubenswrapper[4931]: I1129 01:14:50.044678 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2b4d256-9b86-48e3-98d8-266271a338dd/mysql-bootstrap/0.log" Nov 29 01:14:50 crc kubenswrapper[4931]: I1129 01:14:50.056623 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2b4d256-9b86-48e3-98d8-266271a338dd/galera/0.log" Nov 29 01:14:50 crc kubenswrapper[4931]: I1129 01:14:50.285185 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ba2777f2-23db-46b8-8ccf-cb0813a39678/mysql-bootstrap/0.log" Nov 29 01:14:50 crc kubenswrapper[4931]: I1129 01:14:50.828322 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ba2777f2-23db-46b8-8ccf-cb0813a39678/mysql-bootstrap/0.log" Nov 29 01:14:50 crc kubenswrapper[4931]: I1129 01:14:50.964073 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ba2777f2-23db-46b8-8ccf-cb0813a39678/galera/0.log" Nov 29 01:14:51 crc kubenswrapper[4931]: I1129 01:14:51.028218 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_6440833b-24fe-4da5-afae-6111fc7e760e/nova-metadata-metadata/0.log" Nov 29 01:14:51 crc kubenswrapper[4931]: I1129 01:14:51.029698 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_c9276f25-9d5a-4953-9145-4da299973ff4/openstackclient/0.log" Nov 29 01:14:51 crc kubenswrapper[4931]: I1129 01:14:51.250120 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7s9fx_28c11b0b-3026-4f9f-8de8-71174fbd19d6/ovsdb-server-init/0.log" Nov 29 01:14:51 crc kubenswrapper[4931]: I1129 01:14:51.280095 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-m9dkp_f9b49437-b579-4f61-82a3-57d81e77b0b7/openstack-network-exporter/0.log" Nov 29 01:14:51 crc kubenswrapper[4931]: I1129 01:14:51.477534 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7s9fx_28c11b0b-3026-4f9f-8de8-71174fbd19d6/ovsdb-server/0.log" Nov 29 01:14:51 crc kubenswrapper[4931]: I1129 01:14:51.479006 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7s9fx_28c11b0b-3026-4f9f-8de8-71174fbd19d6/ovsdb-server-init/0.log" Nov 29 01:14:51 crc kubenswrapper[4931]: I1129 01:14:51.483009 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7s9fx_28c11b0b-3026-4f9f-8de8-71174fbd19d6/ovs-vswitchd/0.log" Nov 29 01:14:51 crc kubenswrapper[4931]: I1129 01:14:51.685884 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-trsph_a64bc6bd-8033-45c9-b2d7-9602aea09225/ovn-controller/0.log" Nov 29 01:14:51 crc kubenswrapper[4931]: I1129 01:14:51.756142 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-4p9cv_7715e2f2-11a1-4702-86f0-84744ae25c5d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:51 crc kubenswrapper[4931]: I1129 01:14:51.950640 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb/openstack-network-exporter/0.log" Nov 29 01:14:51 crc kubenswrapper[4931]: I1129 01:14:51.993209 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d7c4acbb-4e74-4467-a524-bc9cf9bb4fdb/ovn-northd/0.log" Nov 29 01:14:52 crc kubenswrapper[4931]: I1129 01:14:52.123023 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_979903af-0c72-45fe-a3de-0dc52370bc57/ovsdbserver-nb/0.log" Nov 29 01:14:52 crc kubenswrapper[4931]: I1129 01:14:52.183668 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_979903af-0c72-45fe-a3de-0dc52370bc57/openstack-network-exporter/0.log" Nov 29 01:14:52 crc kubenswrapper[4931]: I1129 01:14:52.261360 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_37afe9f6-664f-4f51-90a9-485e87a1db92/openstack-network-exporter/0.log" Nov 29 01:14:52 crc kubenswrapper[4931]: I1129 01:14:52.373949 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_37afe9f6-664f-4f51-90a9-485e87a1db92/ovsdbserver-sb/0.log" Nov 29 01:14:52 crc kubenswrapper[4931]: I1129 01:14:52.572681 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6b5869645b-zlntx_48538810-7ec7-4d21-9cf3-1c0ee151eee5/placement-api/0.log" Nov 29 01:14:52 crc kubenswrapper[4931]: I1129 01:14:52.573562 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6b5869645b-zlntx_48538810-7ec7-4d21-9cf3-1c0ee151eee5/placement-log/0.log" Nov 29 01:14:52 crc kubenswrapper[4931]: I1129 01:14:52.584198 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_421b0e7c-7dcd-4e58-a2ee-445c18ec4b22/setup-container/0.log" Nov 29 01:14:52 crc kubenswrapper[4931]: I1129 01:14:52.803168 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_421b0e7c-7dcd-4e58-a2ee-445c18ec4b22/setup-container/0.log" Nov 29 01:14:52 crc kubenswrapper[4931]: I1129 01:14:52.857734 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_421b0e7c-7dcd-4e58-a2ee-445c18ec4b22/rabbitmq/0.log" Nov 29 01:14:52 crc kubenswrapper[4931]: I1129 01:14:52.864992 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f8ccbdb2-db21-4fce-9111-b34a6800e238/setup-container/0.log" Nov 29 01:14:53 crc kubenswrapper[4931]: I1129 01:14:53.055938 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f8ccbdb2-db21-4fce-9111-b34a6800e238/setup-container/0.log" Nov 29 01:14:53 crc kubenswrapper[4931]: I1129 01:14:53.125426 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f8ccbdb2-db21-4fce-9111-b34a6800e238/rabbitmq/0.log" Nov 29 01:14:53 crc kubenswrapper[4931]: I1129 01:14:53.350134 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-gblfs_417e828a-2ea5-4ad0-86e0-b5d612cccbe2/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:53 crc kubenswrapper[4931]: I1129 01:14:53.472864 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-n6rmv_2921aa97-795f-434a-ac6e-009e6555e864/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:53 crc kubenswrapper[4931]: I1129 01:14:53.508249 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-kgqc9_8449a5b7-2fa6-4966-9d34-0243e3536f2b/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:53 crc kubenswrapper[4931]: I1129 01:14:53.669320 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-gfsvs_333e0546-64b0-412e-9059-1e7a4082d360/ssh-known-hosts-edpm-deployment/0.log" Nov 29 01:14:53 crc kubenswrapper[4931]: I1129 01:14:53.690724 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-9gcqp_6c31a7d4-ea5c-4434-a405-cd4148ca5a87/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:53 crc kubenswrapper[4931]: I1129 01:14:53.735394 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 01:14:53 crc kubenswrapper[4931]: I1129 01:14:53.735456 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 01:14:53 crc kubenswrapper[4931]: I1129 01:14:53.923308 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-764cb44577-7q67f_550e16f7-f101-4a91-8e2f-45af3430d6b9/proxy-server/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.056052 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-764cb44577-7q67f_550e16f7-f101-4a91-8e2f-45af3430d6b9/proxy-httpd/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.112408 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/account-auditor/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.161824 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-jnxgv_64db2499-e411-42b9-8e99-92e748a9ad5d/swift-ring-rebalance/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.330751 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/account-reaper/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.343564 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/account-replicator/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.375742 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/account-server/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.408516 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/container-auditor/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.528852 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/container-server/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.585048 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/container-replicator/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.597865 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/container-updater/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.662721 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/object-auditor/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.758557 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/object-expirer/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.796910 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/object-server/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.839460 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/object-replicator/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.897365 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/object-updater/0.log" Nov 29 01:14:54 crc kubenswrapper[4931]: I1129 01:14:54.981144 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/rsync/0.log" Nov 29 01:14:55 crc kubenswrapper[4931]: I1129 01:14:55.001900 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a720d879-8b09-452d-8371-9bf75f28cbde/swift-recon-cron/0.log" Nov 29 01:14:55 crc kubenswrapper[4931]: I1129 01:14:55.186002 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-zvvsj_6a410480-c08c-4a1a-b3a4-6d2372d6bb2a/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:14:55 crc kubenswrapper[4931]: I1129 01:14:55.242374 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_16c927ea-8729-4da8-8893-c23a53d85c5d/tempest-tests-tempest-tests-runner/0.log" Nov 29 01:14:55 crc kubenswrapper[4931]: I1129 01:14:55.341226 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_dbcaae8b-9df7-4dda-a4ea-3a237c20627f/test-operator-logs-container/0.log" Nov 29 01:14:55 crc kubenswrapper[4931]: I1129 01:14:55.484752 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-g26zs_01b89e61-27d6-407a-97e0-7ced1cf90139/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.151871 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf"] Nov 29 01:15:00 crc kubenswrapper[4931]: E1129 01:15:00.152911 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a25d298-51d7-4759-9e62-7157d0e4c584" containerName="container-00" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.152930 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a25d298-51d7-4759-9e62-7157d0e4c584" containerName="container-00" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.153237 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a25d298-51d7-4759-9e62-7157d0e4c584" containerName="container-00" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.154288 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.157208 4931 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.157388 4931 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.182166 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf"] Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.285510 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-config-volume\") pod \"collect-profiles-29406315-sn8kf\" (UID: \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.285582 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ddhq\" (UniqueName: \"kubernetes.io/projected/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-kube-api-access-8ddhq\") pod \"collect-profiles-29406315-sn8kf\" (UID: \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.285644 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-secret-volume\") pod \"collect-profiles-29406315-sn8kf\" (UID: \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.386634 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-secret-volume\") pod \"collect-profiles-29406315-sn8kf\" (UID: \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.386792 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-config-volume\") pod \"collect-profiles-29406315-sn8kf\" (UID: \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.386858 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ddhq\" (UniqueName: \"kubernetes.io/projected/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-kube-api-access-8ddhq\") pod \"collect-profiles-29406315-sn8kf\" (UID: \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.388667 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-config-volume\") pod \"collect-profiles-29406315-sn8kf\" (UID: \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.396298 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-secret-volume\") pod \"collect-profiles-29406315-sn8kf\" (UID: \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.408462 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ddhq\" (UniqueName: \"kubernetes.io/projected/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-kube-api-access-8ddhq\") pod \"collect-profiles-29406315-sn8kf\" (UID: \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:00 crc kubenswrapper[4931]: I1129 01:15:00.479113 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:01 crc kubenswrapper[4931]: I1129 01:15:01.001070 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf"] Nov 29 01:15:01 crc kubenswrapper[4931]: I1129 01:15:01.426943 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" event={"ID":"40ea6e07-ece3-4640-8bee-84b39e2a8fc5","Type":"ContainerStarted","Data":"044bd0ac8088ff8d9bdec50f8ffdbb60920aec424c97a7425efd8f798779e161"} Nov 29 01:15:02 crc kubenswrapper[4931]: I1129 01:15:02.435695 4931 generic.go:334] "Generic (PLEG): container finished" podID="40ea6e07-ece3-4640-8bee-84b39e2a8fc5" containerID="bfddad941a8cd64e6bff7fd8565f564a64652493236b319c50b909ed2a7ed4a6" exitCode=0 Nov 29 01:15:02 crc kubenswrapper[4931]: I1129 01:15:02.436223 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" event={"ID":"40ea6e07-ece3-4640-8bee-84b39e2a8fc5","Type":"ContainerDied","Data":"bfddad941a8cd64e6bff7fd8565f564a64652493236b319c50b909ed2a7ed4a6"} Nov 29 01:15:03 crc kubenswrapper[4931]: I1129 01:15:03.797244 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:03 crc kubenswrapper[4931]: I1129 01:15:03.839077 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ddhq\" (UniqueName: \"kubernetes.io/projected/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-kube-api-access-8ddhq\") pod \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\" (UID: \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\") " Nov 29 01:15:03 crc kubenswrapper[4931]: I1129 01:15:03.841041 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-secret-volume\") pod \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\" (UID: \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\") " Nov 29 01:15:03 crc kubenswrapper[4931]: I1129 01:15:03.841573 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-config-volume\") pod \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\" (UID: \"40ea6e07-ece3-4640-8bee-84b39e2a8fc5\") " Nov 29 01:15:03 crc kubenswrapper[4931]: I1129 01:15:03.851587 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-config-volume" (OuterVolumeSpecName: "config-volume") pod "40ea6e07-ece3-4640-8bee-84b39e2a8fc5" (UID: "40ea6e07-ece3-4640-8bee-84b39e2a8fc5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 29 01:15:03 crc kubenswrapper[4931]: I1129 01:15:03.851922 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "40ea6e07-ece3-4640-8bee-84b39e2a8fc5" (UID: "40ea6e07-ece3-4640-8bee-84b39e2a8fc5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 29 01:15:03 crc kubenswrapper[4931]: I1129 01:15:03.855092 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-kube-api-access-8ddhq" (OuterVolumeSpecName: "kube-api-access-8ddhq") pod "40ea6e07-ece3-4640-8bee-84b39e2a8fc5" (UID: "40ea6e07-ece3-4640-8bee-84b39e2a8fc5"). InnerVolumeSpecName "kube-api-access-8ddhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:15:03 crc kubenswrapper[4931]: I1129 01:15:03.945657 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ddhq\" (UniqueName: \"kubernetes.io/projected/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-kube-api-access-8ddhq\") on node \"crc\" DevicePath \"\"" Nov 29 01:15:03 crc kubenswrapper[4931]: I1129 01:15:03.945871 4931 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 29 01:15:03 crc kubenswrapper[4931]: I1129 01:15:03.945928 4931 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ea6e07-ece3-4640-8bee-84b39e2a8fc5-config-volume\") on node \"crc\" DevicePath \"\"" Nov 29 01:15:04 crc kubenswrapper[4931]: I1129 01:15:04.476240 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" event={"ID":"40ea6e07-ece3-4640-8bee-84b39e2a8fc5","Type":"ContainerDied","Data":"044bd0ac8088ff8d9bdec50f8ffdbb60920aec424c97a7425efd8f798779e161"} Nov 29 01:15:04 crc kubenswrapper[4931]: I1129 01:15:04.476530 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="044bd0ac8088ff8d9bdec50f8ffdbb60920aec424c97a7425efd8f798779e161" Nov 29 01:15:04 crc kubenswrapper[4931]: I1129 01:15:04.476296 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29406315-sn8kf" Nov 29 01:15:04 crc kubenswrapper[4931]: I1129 01:15:04.869310 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8"] Nov 29 01:15:04 crc kubenswrapper[4931]: I1129 01:15:04.880378 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29406270-l8vn8"] Nov 29 01:15:05 crc kubenswrapper[4931]: I1129 01:15:05.034980 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_d60ec023-b0d6-466c-b188-5e0e2b47a962/memcached/0.log" Nov 29 01:15:05 crc kubenswrapper[4931]: I1129 01:15:05.222776 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6" path="/var/lib/kubelet/pods/489ce73e-e95a-4f83-9a42-0e9ffcdfc2b6/volumes" Nov 29 01:15:17 crc kubenswrapper[4931]: I1129 01:15:17.974178 4931 scope.go:117] "RemoveContainer" containerID="7884d5457e9b5abc28e60b8ed77e2eb30fb101add86ac5f94b1002c5faf2c62a" Nov 29 01:15:23 crc kubenswrapper[4931]: I1129 01:15:23.556892 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/util/0.log" Nov 29 01:15:23 crc kubenswrapper[4931]: I1129 01:15:23.735043 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 01:15:23 crc kubenswrapper[4931]: I1129 01:15:23.735138 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 01:15:23 crc kubenswrapper[4931]: I1129 01:15:23.767142 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/pull/0.log" Nov 29 01:15:23 crc kubenswrapper[4931]: I1129 01:15:23.773654 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/util/0.log" Nov 29 01:15:23 crc kubenswrapper[4931]: I1129 01:15:23.788599 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/pull/0.log" Nov 29 01:15:23 crc kubenswrapper[4931]: I1129 01:15:23.973850 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/util/0.log" Nov 29 01:15:23 crc kubenswrapper[4931]: I1129 01:15:23.985649 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/extract/0.log" Nov 29 01:15:23 crc kubenswrapper[4931]: I1129 01:15:23.993852 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_287aedcf25cf3b1681db73853e637b3557269d88c4e881b8d9ff773c1ehbl2w_ef059f7c-88cc-442c-9f6e-5e7a5778a549/pull/0.log" Nov 29 01:15:24 crc kubenswrapper[4931]: I1129 01:15:24.160571 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-qkqqf_64f054e4-1ff5-4d20-a098-499b8baa25eb/kube-rbac-proxy/0.log" Nov 29 01:15:24 crc kubenswrapper[4931]: I1129 01:15:24.231502 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-qkqqf_64f054e4-1ff5-4d20-a098-499b8baa25eb/manager/0.log" Nov 29 01:15:24 crc kubenswrapper[4931]: I1129 01:15:24.331841 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-tqw4f_551c1c83-9dd7-47c7-8411-30881adc79bb/kube-rbac-proxy/0.log" Nov 29 01:15:24 crc kubenswrapper[4931]: I1129 01:15:24.399533 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-tqw4f_551c1c83-9dd7-47c7-8411-30881adc79bb/manager/0.log" Nov 29 01:15:24 crc kubenswrapper[4931]: I1129 01:15:24.453921 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-wzvc6_256e71e1-5f65-4842-883e-1e3178e91965/kube-rbac-proxy/0.log" Nov 29 01:15:24 crc kubenswrapper[4931]: I1129 01:15:24.555374 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-wzvc6_256e71e1-5f65-4842-883e-1e3178e91965/manager/0.log" Nov 29 01:15:24 crc kubenswrapper[4931]: I1129 01:15:24.666466 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-lhc87_b870bbdd-21a8-4279-9d2a-12dd26745325/kube-rbac-proxy/0.log" Nov 29 01:15:24 crc kubenswrapper[4931]: I1129 01:15:24.782948 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-lhc87_b870bbdd-21a8-4279-9d2a-12dd26745325/manager/0.log" Nov 29 01:15:24 crc kubenswrapper[4931]: I1129 01:15:24.884305 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-nmtz7_21842304-a783-4f75-bbb5-2ece9dacb6ca/manager/0.log" Nov 29 01:15:24 crc kubenswrapper[4931]: I1129 01:15:24.908097 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-nmtz7_21842304-a783-4f75-bbb5-2ece9dacb6ca/kube-rbac-proxy/0.log" Nov 29 01:15:25 crc kubenswrapper[4931]: I1129 01:15:25.088526 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-r5d5m_6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c/kube-rbac-proxy/0.log" Nov 29 01:15:25 crc kubenswrapper[4931]: I1129 01:15:25.123027 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-r5d5m_6c3cd3e3-0977-4a0d-8a36-35bfe26ab12c/manager/0.log" Nov 29 01:15:25 crc kubenswrapper[4931]: I1129 01:15:25.267367 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-nkmhr_3a157886-d4b5-4019-a0d6-0b40f6509b21/kube-rbac-proxy/0.log" Nov 29 01:15:25 crc kubenswrapper[4931]: I1129 01:15:25.350492 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-zb94f_5f790baf-d77a-41dd-840c-f4d9e13420e1/kube-rbac-proxy/0.log" Nov 29 01:15:25 crc kubenswrapper[4931]: I1129 01:15:25.461696 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-nkmhr_3a157886-d4b5-4019-a0d6-0b40f6509b21/manager/0.log" Nov 29 01:15:25 crc kubenswrapper[4931]: I1129 01:15:25.484093 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-zb94f_5f790baf-d77a-41dd-840c-f4d9e13420e1/manager/0.log" Nov 29 01:15:25 crc kubenswrapper[4931]: I1129 01:15:25.737175 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-g82st_0ce6e577-25c0-4aee-b14f-47e02ebfbef9/kube-rbac-proxy/0.log" Nov 29 01:15:25 crc kubenswrapper[4931]: I1129 01:15:25.885995 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-g82st_0ce6e577-25c0-4aee-b14f-47e02ebfbef9/manager/0.log" Nov 29 01:15:25 crc kubenswrapper[4931]: I1129 01:15:25.940868 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-b4skl_5a12e787-bf00-4f2b-8760-9b9a5cad0834/kube-rbac-proxy/0.log" Nov 29 01:15:26 crc kubenswrapper[4931]: I1129 01:15:26.075876 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-b4skl_5a12e787-bf00-4f2b-8760-9b9a5cad0834/manager/0.log" Nov 29 01:15:26 crc kubenswrapper[4931]: I1129 01:15:26.107138 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-2dbgl_724defda-f0a4-4edd-8a1a-7f6625893fe7/kube-rbac-proxy/0.log" Nov 29 01:15:26 crc kubenswrapper[4931]: I1129 01:15:26.216931 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-2dbgl_724defda-f0a4-4edd-8a1a-7f6625893fe7/manager/0.log" Nov 29 01:15:27 crc kubenswrapper[4931]: I1129 01:15:27.029440 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-q7q7q_365eb479-0132-4a0c-99d3-01cb923a9d61/manager/0.log" Nov 29 01:15:27 crc kubenswrapper[4931]: I1129 01:15:27.050396 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-q7q7q_365eb479-0132-4a0c-99d3-01cb923a9d61/kube-rbac-proxy/0.log" Nov 29 01:15:27 crc kubenswrapper[4931]: I1129 01:15:27.182188 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-lpf99_c352f7a8-72dc-41da-9673-7fccd1a0974f/kube-rbac-proxy/0.log" Nov 29 01:15:27 crc kubenswrapper[4931]: I1129 01:15:27.274921 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-lpf99_c352f7a8-72dc-41da-9673-7fccd1a0974f/manager/0.log" Nov 29 01:15:27 crc kubenswrapper[4931]: I1129 01:15:27.332837 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-z2mjc_4c70c5cd-a329-4c90-b454-f836ddaef38f/manager/0.log" Nov 29 01:15:27 crc kubenswrapper[4931]: I1129 01:15:27.336435 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-z2mjc_4c70c5cd-a329-4c90-b454-f836ddaef38f/kube-rbac-proxy/0.log" Nov 29 01:15:27 crc kubenswrapper[4931]: I1129 01:15:27.473205 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp_fd6606d5-8cd5-4308-a010-eea75cb666f1/kube-rbac-proxy/0.log" Nov 29 01:15:27 crc kubenswrapper[4931]: I1129 01:15:27.509603 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4x6fdp_fd6606d5-8cd5-4308-a010-eea75cb666f1/manager/0.log" Nov 29 01:15:27 crc kubenswrapper[4931]: I1129 01:15:27.815426 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-h7scz_4ef2db4a-ba6d-47a1-8054-b61f854da2d0/registry-server/0.log" Nov 29 01:15:27 crc kubenswrapper[4931]: I1129 01:15:27.892363 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5757985b9c-rwxqq_fe758412-e16a-4c07-8c0d-e79e9f54cd73/operator/0.log" Nov 29 01:15:28 crc kubenswrapper[4931]: I1129 01:15:28.059367 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lnrrq_6f4cd8c8-caf0-4191-a3c4-51a172812b9b/kube-rbac-proxy/0.log" Nov 29 01:15:28 crc kubenswrapper[4931]: I1129 01:15:28.062408 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lnrrq_6f4cd8c8-caf0-4191-a3c4-51a172812b9b/manager/0.log" Nov 29 01:15:28 crc kubenswrapper[4931]: I1129 01:15:28.142454 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-lc47s_96c0b73e-c18e-4a52-a20c-e10db096b799/kube-rbac-proxy/0.log" Nov 29 01:15:28 crc kubenswrapper[4931]: I1129 01:15:28.285210 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-lc47s_96c0b73e-c18e-4a52-a20c-e10db096b799/manager/0.log" Nov 29 01:15:28 crc kubenswrapper[4931]: I1129 01:15:28.356481 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-4qjb4_22500bd3-4958-4ba1-9b59-deb3032aea18/operator/0.log" Nov 29 01:15:28 crc kubenswrapper[4931]: I1129 01:15:28.700114 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74fb9455b7-sz2k9_0c61786d-9cf4-4447-82f7-47cc17ac65d8/manager/0.log" Nov 29 01:15:28 crc kubenswrapper[4931]: I1129 01:15:28.933897 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-f62b4_249844ae-a64c-4f9d-b204-bf3a1956aaa7/kube-rbac-proxy/0.log" Nov 29 01:15:28 crc kubenswrapper[4931]: I1129 01:15:28.974006 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-f62b4_249844ae-a64c-4f9d-b204-bf3a1956aaa7/manager/0.log" Nov 29 01:15:28 crc kubenswrapper[4931]: I1129 01:15:28.992937 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-lthd8_b57413c5-3888-496c-ad0b-18128019b945/kube-rbac-proxy/0.log" Nov 29 01:15:29 crc kubenswrapper[4931]: I1129 01:15:29.170326 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-lthd8_b57413c5-3888-496c-ad0b-18128019b945/manager/0.log" Nov 29 01:15:29 crc kubenswrapper[4931]: I1129 01:15:29.174553 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-dcl69_9621ec40-d523-423e-9bbe-4b063355cf08/kube-rbac-proxy/0.log" Nov 29 01:15:29 crc kubenswrapper[4931]: I1129 01:15:29.199990 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-dcl69_9621ec40-d523-423e-9bbe-4b063355cf08/manager/0.log" Nov 29 01:15:29 crc kubenswrapper[4931]: I1129 01:15:29.309942 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-qp9ht_2792db7c-cdd1-4356-97d5-fdba06bd44e4/kube-rbac-proxy/0.log" Nov 29 01:15:29 crc kubenswrapper[4931]: I1129 01:15:29.355995 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-qp9ht_2792db7c-cdd1-4356-97d5-fdba06bd44e4/manager/0.log" Nov 29 01:15:49 crc kubenswrapper[4931]: I1129 01:15:49.194334 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-pgslt_b6c59839-1a09-4b0b-90ac-eb9d001e9581/control-plane-machine-set-operator/0.log" Nov 29 01:15:49 crc kubenswrapper[4931]: I1129 01:15:49.363390 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xl2bw_980879e9-818d-414c-97c0-ad63e6cf4039/kube-rbac-proxy/0.log" Nov 29 01:15:49 crc kubenswrapper[4931]: I1129 01:15:49.398606 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xl2bw_980879e9-818d-414c-97c0-ad63e6cf4039/machine-api-operator/0.log" Nov 29 01:15:53 crc kubenswrapper[4931]: I1129 01:15:53.734697 4931 patch_prober.go:28] interesting pod/machine-config-daemon-6xxhj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 29 01:15:53 crc kubenswrapper[4931]: I1129 01:15:53.735321 4931 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 29 01:15:53 crc kubenswrapper[4931]: I1129 01:15:53.735379 4931 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" Nov 29 01:15:53 crc kubenswrapper[4931]: I1129 01:15:53.736248 4931 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a"} pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 29 01:15:53 crc kubenswrapper[4931]: I1129 01:15:53.736317 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" containerName="machine-config-daemon" containerID="cri-o://11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" gracePeriod=600 Nov 29 01:15:53 crc kubenswrapper[4931]: E1129 01:15:53.861036 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:15:53 crc kubenswrapper[4931]: I1129 01:15:53.949018 4931 generic.go:334] "Generic (PLEG): container finished" podID="0b227491-f635-4817-96ed-90b96ab6aab3" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" exitCode=0 Nov 29 01:15:53 crc kubenswrapper[4931]: I1129 01:15:53.949084 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerDied","Data":"11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a"} Nov 29 01:15:53 crc kubenswrapper[4931]: I1129 01:15:53.949151 4931 scope.go:117] "RemoveContainer" containerID="095e15e06fb756ce4c0646d5a8622ecc181be42bd4003d1af6ab96ca0122a6bd" Nov 29 01:15:53 crc kubenswrapper[4931]: I1129 01:15:53.949847 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:15:53 crc kubenswrapper[4931]: E1129 01:15:53.950194 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:16:02 crc kubenswrapper[4931]: I1129 01:16:02.888287 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-s9fhh_ea00643b-3840-4675-b14a-323b377e2339/cert-manager-controller/0.log" Nov 29 01:16:03 crc kubenswrapper[4931]: I1129 01:16:03.014034 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-b8cd4_283e1d51-cb76-418b-b4b0-d444320900ff/cert-manager-webhook/0.log" Nov 29 01:16:03 crc kubenswrapper[4931]: I1129 01:16:03.036974 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-g6bkx_2f242144-c19c-4886-a51b-8547f684e2e9/cert-manager-cainjector/0.log" Nov 29 01:16:07 crc kubenswrapper[4931]: I1129 01:16:07.220634 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:16:07 crc kubenswrapper[4931]: E1129 01:16:07.222631 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:16:11 crc kubenswrapper[4931]: I1129 01:16:11.971972 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2s2tx"] Nov 29 01:16:11 crc kubenswrapper[4931]: E1129 01:16:11.972775 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40ea6e07-ece3-4640-8bee-84b39e2a8fc5" containerName="collect-profiles" Nov 29 01:16:11 crc kubenswrapper[4931]: I1129 01:16:11.972789 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="40ea6e07-ece3-4640-8bee-84b39e2a8fc5" containerName="collect-profiles" Nov 29 01:16:11 crc kubenswrapper[4931]: I1129 01:16:11.973087 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="40ea6e07-ece3-4640-8bee-84b39e2a8fc5" containerName="collect-profiles" Nov 29 01:16:11 crc kubenswrapper[4931]: I1129 01:16:11.974482 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:11 crc kubenswrapper[4931]: I1129 01:16:11.984137 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2s2tx"] Nov 29 01:16:12 crc kubenswrapper[4931]: I1129 01:16:12.072607 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-catalog-content\") pod \"redhat-marketplace-2s2tx\" (UID: \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\") " pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:12 crc kubenswrapper[4931]: I1129 01:16:12.072694 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72w6b\" (UniqueName: \"kubernetes.io/projected/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-kube-api-access-72w6b\") pod \"redhat-marketplace-2s2tx\" (UID: \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\") " pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:12 crc kubenswrapper[4931]: I1129 01:16:12.072781 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-utilities\") pod \"redhat-marketplace-2s2tx\" (UID: \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\") " pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:12 crc kubenswrapper[4931]: I1129 01:16:12.174306 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-catalog-content\") pod \"redhat-marketplace-2s2tx\" (UID: \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\") " pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:12 crc kubenswrapper[4931]: I1129 01:16:12.174362 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72w6b\" (UniqueName: \"kubernetes.io/projected/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-kube-api-access-72w6b\") pod \"redhat-marketplace-2s2tx\" (UID: \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\") " pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:12 crc kubenswrapper[4931]: I1129 01:16:12.174450 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-utilities\") pod \"redhat-marketplace-2s2tx\" (UID: \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\") " pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:12 crc kubenswrapper[4931]: I1129 01:16:12.174886 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-catalog-content\") pod \"redhat-marketplace-2s2tx\" (UID: \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\") " pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:12 crc kubenswrapper[4931]: I1129 01:16:12.174943 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-utilities\") pod \"redhat-marketplace-2s2tx\" (UID: \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\") " pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:12 crc kubenswrapper[4931]: I1129 01:16:12.194768 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72w6b\" (UniqueName: \"kubernetes.io/projected/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-kube-api-access-72w6b\") pod \"redhat-marketplace-2s2tx\" (UID: \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\") " pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:12 crc kubenswrapper[4931]: I1129 01:16:12.303657 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:12 crc kubenswrapper[4931]: I1129 01:16:12.756674 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2s2tx"] Nov 29 01:16:13 crc kubenswrapper[4931]: I1129 01:16:13.137008 4931 generic.go:334] "Generic (PLEG): container finished" podID="fe6655fe-6ac3-415f-a87d-6a85766fc3fd" containerID="fd98ea0ba8b220f58de3d63d68c300d9807fd31235276c4a1c7841a77eb2c674" exitCode=0 Nov 29 01:16:13 crc kubenswrapper[4931]: I1129 01:16:13.137053 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2s2tx" event={"ID":"fe6655fe-6ac3-415f-a87d-6a85766fc3fd","Type":"ContainerDied","Data":"fd98ea0ba8b220f58de3d63d68c300d9807fd31235276c4a1c7841a77eb2c674"} Nov 29 01:16:13 crc kubenswrapper[4931]: I1129 01:16:13.137079 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2s2tx" event={"ID":"fe6655fe-6ac3-415f-a87d-6a85766fc3fd","Type":"ContainerStarted","Data":"638dd99382411f71863d477d792521334550ba7ec89ba29784f88a4108971444"} Nov 29 01:16:13 crc kubenswrapper[4931]: I1129 01:16:13.139139 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 01:16:13 crc kubenswrapper[4931]: I1129 01:16:13.982472 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hjzww"] Nov 29 01:16:13 crc kubenswrapper[4931]: I1129 01:16:13.987830 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:14 crc kubenswrapper[4931]: I1129 01:16:14.040867 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hjzww"] Nov 29 01:16:14 crc kubenswrapper[4931]: I1129 01:16:14.110236 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-catalog-content\") pod \"redhat-operators-hjzww\" (UID: \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\") " pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:14 crc kubenswrapper[4931]: I1129 01:16:14.110526 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-utilities\") pod \"redhat-operators-hjzww\" (UID: \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\") " pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:14 crc kubenswrapper[4931]: I1129 01:16:14.110578 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd4tn\" (UniqueName: \"kubernetes.io/projected/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-kube-api-access-jd4tn\") pod \"redhat-operators-hjzww\" (UID: \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\") " pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:14 crc kubenswrapper[4931]: I1129 01:16:14.214918 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-catalog-content\") pod \"redhat-operators-hjzww\" (UID: \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\") " pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:14 crc kubenswrapper[4931]: I1129 01:16:14.214996 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-utilities\") pod \"redhat-operators-hjzww\" (UID: \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\") " pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:14 crc kubenswrapper[4931]: I1129 01:16:14.215061 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd4tn\" (UniqueName: \"kubernetes.io/projected/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-kube-api-access-jd4tn\") pod \"redhat-operators-hjzww\" (UID: \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\") " pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:14 crc kubenswrapper[4931]: I1129 01:16:14.215715 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-catalog-content\") pod \"redhat-operators-hjzww\" (UID: \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\") " pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:14 crc kubenswrapper[4931]: I1129 01:16:14.215941 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-utilities\") pod \"redhat-operators-hjzww\" (UID: \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\") " pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:14 crc kubenswrapper[4931]: I1129 01:16:14.235264 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd4tn\" (UniqueName: \"kubernetes.io/projected/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-kube-api-access-jd4tn\") pod \"redhat-operators-hjzww\" (UID: \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\") " pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:14 crc kubenswrapper[4931]: I1129 01:16:14.387796 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:14 crc kubenswrapper[4931]: I1129 01:16:14.838410 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hjzww"] Nov 29 01:16:15 crc kubenswrapper[4931]: I1129 01:16:15.157395 4931 generic.go:334] "Generic (PLEG): container finished" podID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" containerID="71293f474e38d99d0fa4d588bd729a6793948fd542da05f4267e14ba869b6c8c" exitCode=0 Nov 29 01:16:15 crc kubenswrapper[4931]: I1129 01:16:15.157767 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjzww" event={"ID":"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347","Type":"ContainerDied","Data":"71293f474e38d99d0fa4d588bd729a6793948fd542da05f4267e14ba869b6c8c"} Nov 29 01:16:15 crc kubenswrapper[4931]: I1129 01:16:15.157817 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjzww" event={"ID":"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347","Type":"ContainerStarted","Data":"59c8cf6e2dc07f2209353ca60db3023e1421bf89a29656226fc7c9f5103657e0"} Nov 29 01:16:15 crc kubenswrapper[4931]: I1129 01:16:15.161339 4931 generic.go:334] "Generic (PLEG): container finished" podID="fe6655fe-6ac3-415f-a87d-6a85766fc3fd" containerID="1142ee7d46af69c07bf1610c2fd225018495c6fe41d50cd2d1e681471585c2d9" exitCode=0 Nov 29 01:16:15 crc kubenswrapper[4931]: I1129 01:16:15.161377 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2s2tx" event={"ID":"fe6655fe-6ac3-415f-a87d-6a85766fc3fd","Type":"ContainerDied","Data":"1142ee7d46af69c07bf1610c2fd225018495c6fe41d50cd2d1e681471585c2d9"} Nov 29 01:16:17 crc kubenswrapper[4931]: I1129 01:16:17.182273 4931 generic.go:334] "Generic (PLEG): container finished" podID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" containerID="bef9ee83f4f2be5571f73f631bab1f388fe681a54d0a098984c6688b41dd2a25" exitCode=0 Nov 29 01:16:17 crc kubenswrapper[4931]: I1129 01:16:17.182486 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjzww" event={"ID":"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347","Type":"ContainerDied","Data":"bef9ee83f4f2be5571f73f631bab1f388fe681a54d0a098984c6688b41dd2a25"} Nov 29 01:16:17 crc kubenswrapper[4931]: I1129 01:16:17.185765 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2s2tx" event={"ID":"fe6655fe-6ac3-415f-a87d-6a85766fc3fd","Type":"ContainerStarted","Data":"5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618"} Nov 29 01:16:17 crc kubenswrapper[4931]: I1129 01:16:17.262559 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2s2tx" podStartSLOduration=3.60475373 podStartE2EDuration="6.262523775s" podCreationTimestamp="2025-11-29 01:16:11 +0000 UTC" firstStartedPulling="2025-11-29 01:16:13.138937078 +0000 UTC m=+4586.300830310" lastFinishedPulling="2025-11-29 01:16:15.796707113 +0000 UTC m=+4588.958600355" observedRunningTime="2025-11-29 01:16:17.248697496 +0000 UTC m=+4590.410590728" watchObservedRunningTime="2025-11-29 01:16:17.262523775 +0000 UTC m=+4590.424417007" Nov 29 01:16:18 crc kubenswrapper[4931]: I1129 01:16:18.689618 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-g8dpm_6fbf3473-ee9a-47b5-a782-47948ff5e590/nmstate-console-plugin/0.log" Nov 29 01:16:18 crc kubenswrapper[4931]: I1129 01:16:18.879791 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-wv5fd_7622d8c6-7baa-47e7-8f03-52ccd0d2f93f/nmstate-handler/0.log" Nov 29 01:16:19 crc kubenswrapper[4931]: I1129 01:16:19.014669 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-cmcf4_e84ff97d-2d9a-4a4a-b619-a26e9c5a5595/kube-rbac-proxy/0.log" Nov 29 01:16:19 crc kubenswrapper[4931]: I1129 01:16:19.093620 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-t6vpl_3f3c7035-b1e3-48bf-bbb9-3c6400ffedae/nmstate-operator/0.log" Nov 29 01:16:19 crc kubenswrapper[4931]: I1129 01:16:19.160989 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-cmcf4_e84ff97d-2d9a-4a4a-b619-a26e9c5a5595/nmstate-metrics/0.log" Nov 29 01:16:19 crc kubenswrapper[4931]: I1129 01:16:19.206263 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjzww" event={"ID":"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347","Type":"ContainerStarted","Data":"c764a3dd78de84f71cf1a2682d0c9b42c7a8910b7e8a98eb0184cb523d9189c9"} Nov 29 01:16:19 crc kubenswrapper[4931]: I1129 01:16:19.227698 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hjzww" podStartSLOduration=2.920788396 podStartE2EDuration="6.227683333s" podCreationTimestamp="2025-11-29 01:16:13 +0000 UTC" firstStartedPulling="2025-11-29 01:16:15.15895032 +0000 UTC m=+4588.320843562" lastFinishedPulling="2025-11-29 01:16:18.465845267 +0000 UTC m=+4591.627738499" observedRunningTime="2025-11-29 01:16:19.226203021 +0000 UTC m=+4592.388096273" watchObservedRunningTime="2025-11-29 01:16:19.227683333 +0000 UTC m=+4592.389576555" Nov 29 01:16:19 crc kubenswrapper[4931]: I1129 01:16:19.287181 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-jh9h6_4d93b5c8-a612-45f9-9351-31ffc5eb3b5a/nmstate-webhook/0.log" Nov 29 01:16:20 crc kubenswrapper[4931]: I1129 01:16:20.215494 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:16:20 crc kubenswrapper[4931]: E1129 01:16:20.216013 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:16:22 crc kubenswrapper[4931]: I1129 01:16:22.304041 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:22 crc kubenswrapper[4931]: I1129 01:16:22.304294 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:22 crc kubenswrapper[4931]: I1129 01:16:22.354719 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:23 crc kubenswrapper[4931]: I1129 01:16:23.287412 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:23 crc kubenswrapper[4931]: I1129 01:16:23.961403 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2s2tx"] Nov 29 01:16:24 crc kubenswrapper[4931]: I1129 01:16:24.388064 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:24 crc kubenswrapper[4931]: I1129 01:16:24.388183 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:25 crc kubenswrapper[4931]: I1129 01:16:25.258059 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2s2tx" podUID="fe6655fe-6ac3-415f-a87d-6a85766fc3fd" containerName="registry-server" containerID="cri-o://5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618" gracePeriod=2 Nov 29 01:16:25 crc kubenswrapper[4931]: I1129 01:16:25.441464 4931 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hjzww" podUID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" containerName="registry-server" probeResult="failure" output=< Nov 29 01:16:25 crc kubenswrapper[4931]: timeout: failed to connect service ":50051" within 1s Nov 29 01:16:25 crc kubenswrapper[4931]: > Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.273744 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.279511 4931 generic.go:334] "Generic (PLEG): container finished" podID="fe6655fe-6ac3-415f-a87d-6a85766fc3fd" containerID="5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618" exitCode=0 Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.279543 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2s2tx" event={"ID":"fe6655fe-6ac3-415f-a87d-6a85766fc3fd","Type":"ContainerDied","Data":"5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618"} Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.280676 4931 scope.go:117] "RemoveContainer" containerID="5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.280691 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2s2tx" event={"ID":"fe6655fe-6ac3-415f-a87d-6a85766fc3fd","Type":"ContainerDied","Data":"638dd99382411f71863d477d792521334550ba7ec89ba29784f88a4108971444"} Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.324140 4931 scope.go:117] "RemoveContainer" containerID="1142ee7d46af69c07bf1610c2fd225018495c6fe41d50cd2d1e681471585c2d9" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.339237 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72w6b\" (UniqueName: \"kubernetes.io/projected/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-kube-api-access-72w6b\") pod \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\" (UID: \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\") " Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.339550 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-utilities\") pod \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\" (UID: \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\") " Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.339640 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-catalog-content\") pod \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\" (UID: \"fe6655fe-6ac3-415f-a87d-6a85766fc3fd\") " Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.343940 4931 scope.go:117] "RemoveContainer" containerID="fd98ea0ba8b220f58de3d63d68c300d9807fd31235276c4a1c7841a77eb2c674" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.344960 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-utilities" (OuterVolumeSpecName: "utilities") pod "fe6655fe-6ac3-415f-a87d-6a85766fc3fd" (UID: "fe6655fe-6ac3-415f-a87d-6a85766fc3fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.345958 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-kube-api-access-72w6b" (OuterVolumeSpecName: "kube-api-access-72w6b") pod "fe6655fe-6ac3-415f-a87d-6a85766fc3fd" (UID: "fe6655fe-6ac3-415f-a87d-6a85766fc3fd"). InnerVolumeSpecName "kube-api-access-72w6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.363270 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe6655fe-6ac3-415f-a87d-6a85766fc3fd" (UID: "fe6655fe-6ac3-415f-a87d-6a85766fc3fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.439159 4931 scope.go:117] "RemoveContainer" containerID="5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618" Nov 29 01:16:26 crc kubenswrapper[4931]: E1129 01:16:26.444635 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618\": container with ID starting with 5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618 not found: ID does not exist" containerID="5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.446192 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618"} err="failed to get container status \"5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618\": rpc error: code = NotFound desc = could not find container \"5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618\": container with ID starting with 5593999fec77903c55ed360a7b620c206e195ea3d49c83d2d4b4c3567f354618 not found: ID does not exist" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.446334 4931 scope.go:117] "RemoveContainer" containerID="1142ee7d46af69c07bf1610c2fd225018495c6fe41d50cd2d1e681471585c2d9" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.444925 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72w6b\" (UniqueName: \"kubernetes.io/projected/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-kube-api-access-72w6b\") on node \"crc\" DevicePath \"\"" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.446514 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.446575 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe6655fe-6ac3-415f-a87d-6a85766fc3fd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 01:16:26 crc kubenswrapper[4931]: E1129 01:16:26.448752 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1142ee7d46af69c07bf1610c2fd225018495c6fe41d50cd2d1e681471585c2d9\": container with ID starting with 1142ee7d46af69c07bf1610c2fd225018495c6fe41d50cd2d1e681471585c2d9 not found: ID does not exist" containerID="1142ee7d46af69c07bf1610c2fd225018495c6fe41d50cd2d1e681471585c2d9" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.448788 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1142ee7d46af69c07bf1610c2fd225018495c6fe41d50cd2d1e681471585c2d9"} err="failed to get container status \"1142ee7d46af69c07bf1610c2fd225018495c6fe41d50cd2d1e681471585c2d9\": rpc error: code = NotFound desc = could not find container \"1142ee7d46af69c07bf1610c2fd225018495c6fe41d50cd2d1e681471585c2d9\": container with ID starting with 1142ee7d46af69c07bf1610c2fd225018495c6fe41d50cd2d1e681471585c2d9 not found: ID does not exist" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.448824 4931 scope.go:117] "RemoveContainer" containerID="fd98ea0ba8b220f58de3d63d68c300d9807fd31235276c4a1c7841a77eb2c674" Nov 29 01:16:26 crc kubenswrapper[4931]: E1129 01:16:26.452886 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd98ea0ba8b220f58de3d63d68c300d9807fd31235276c4a1c7841a77eb2c674\": container with ID starting with fd98ea0ba8b220f58de3d63d68c300d9807fd31235276c4a1c7841a77eb2c674 not found: ID does not exist" containerID="fd98ea0ba8b220f58de3d63d68c300d9807fd31235276c4a1c7841a77eb2c674" Nov 29 01:16:26 crc kubenswrapper[4931]: I1129 01:16:26.452914 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd98ea0ba8b220f58de3d63d68c300d9807fd31235276c4a1c7841a77eb2c674"} err="failed to get container status \"fd98ea0ba8b220f58de3d63d68c300d9807fd31235276c4a1c7841a77eb2c674\": rpc error: code = NotFound desc = could not find container \"fd98ea0ba8b220f58de3d63d68c300d9807fd31235276c4a1c7841a77eb2c674\": container with ID starting with fd98ea0ba8b220f58de3d63d68c300d9807fd31235276c4a1c7841a77eb2c674 not found: ID does not exist" Nov 29 01:16:27 crc kubenswrapper[4931]: I1129 01:16:27.289781 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2s2tx" Nov 29 01:16:27 crc kubenswrapper[4931]: I1129 01:16:27.316766 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2s2tx"] Nov 29 01:16:27 crc kubenswrapper[4931]: I1129 01:16:27.326373 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2s2tx"] Nov 29 01:16:29 crc kubenswrapper[4931]: I1129 01:16:29.225876 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe6655fe-6ac3-415f-a87d-6a85766fc3fd" path="/var/lib/kubelet/pods/fe6655fe-6ac3-415f-a87d-6a85766fc3fd/volumes" Nov 29 01:16:34 crc kubenswrapper[4931]: I1129 01:16:34.212833 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:16:34 crc kubenswrapper[4931]: E1129 01:16:34.213539 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:16:34 crc kubenswrapper[4931]: I1129 01:16:34.426669 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-qzb8h_a44839c5-bede-4c88-b3ff-75f0d06074a6/kube-rbac-proxy/0.log" Nov 29 01:16:34 crc kubenswrapper[4931]: I1129 01:16:34.446881 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-qzb8h_a44839c5-bede-4c88-b3ff-75f0d06074a6/controller/0.log" Nov 29 01:16:34 crc kubenswrapper[4931]: I1129 01:16:34.453531 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:34 crc kubenswrapper[4931]: I1129 01:16:34.499346 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:34 crc kubenswrapper[4931]: I1129 01:16:34.607167 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-frr-files/0.log" Nov 29 01:16:34 crc kubenswrapper[4931]: I1129 01:16:34.690349 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hjzww"] Nov 29 01:16:34 crc kubenswrapper[4931]: I1129 01:16:34.747629 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-frr-files/0.log" Nov 29 01:16:34 crc kubenswrapper[4931]: I1129 01:16:34.783690 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-reloader/0.log" Nov 29 01:16:34 crc kubenswrapper[4931]: I1129 01:16:34.801147 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-metrics/0.log" Nov 29 01:16:34 crc kubenswrapper[4931]: I1129 01:16:34.811532 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-reloader/0.log" Nov 29 01:16:34 crc kubenswrapper[4931]: I1129 01:16:34.948507 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-frr-files/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.005078 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-metrics/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.005103 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-reloader/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.031128 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-metrics/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.157080 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-frr-files/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.158174 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-metrics/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.164035 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/cp-reloader/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.207424 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/controller/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.343452 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/kube-rbac-proxy/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.346059 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/frr-metrics/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.399763 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/kube-rbac-proxy-frr/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.566433 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/reloader/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.614224 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-c96lv_fe08c847-a4ed-4e65-a3cb-e2145c69365f/frr-k8s-webhook-server/0.log" Nov 29 01:16:35 crc kubenswrapper[4931]: I1129 01:16:35.793910 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-df4f9f548-5vp72_70a74ef6-a92b-49b3-b533-17b6a15f513c/manager/0.log" Nov 29 01:16:36 crc kubenswrapper[4931]: I1129 01:16:36.013895 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6ck62_2eb6f229-405c-40db-8504-8fb851cb98d4/kube-rbac-proxy/0.log" Nov 29 01:16:36 crc kubenswrapper[4931]: I1129 01:16:36.026601 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-d7b5f7bd5-t4z72_04d440db-7f7e-46fa-9654-58b389ea3ad7/webhook-server/0.log" Nov 29 01:16:36 crc kubenswrapper[4931]: I1129 01:16:36.382609 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hjzww" podUID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" containerName="registry-server" containerID="cri-o://c764a3dd78de84f71cf1a2682d0c9b42c7a8910b7e8a98eb0184cb523d9189c9" gracePeriod=2 Nov 29 01:16:36 crc kubenswrapper[4931]: I1129 01:16:36.712959 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6ck62_2eb6f229-405c-40db-8504-8fb851cb98d4/speaker/0.log" Nov 29 01:16:36 crc kubenswrapper[4931]: I1129 01:16:36.839920 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dff5k_65f9ee6d-ee55-45a6-abaf-22b0c059aad6/frr/0.log" Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.391330 4931 generic.go:334] "Generic (PLEG): container finished" podID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" containerID="c764a3dd78de84f71cf1a2682d0c9b42c7a8910b7e8a98eb0184cb523d9189c9" exitCode=0 Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.391412 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjzww" event={"ID":"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347","Type":"ContainerDied","Data":"c764a3dd78de84f71cf1a2682d0c9b42c7a8910b7e8a98eb0184cb523d9189c9"} Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.391741 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hjzww" event={"ID":"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347","Type":"ContainerDied","Data":"59c8cf6e2dc07f2209353ca60db3023e1421bf89a29656226fc7c9f5103657e0"} Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.391767 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59c8cf6e2dc07f2209353ca60db3023e1421bf89a29656226fc7c9f5103657e0" Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.401666 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.468305 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd4tn\" (UniqueName: \"kubernetes.io/projected/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-kube-api-access-jd4tn\") pod \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\" (UID: \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\") " Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.468406 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-utilities\") pod \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\" (UID: \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\") " Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.468464 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-catalog-content\") pod \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\" (UID: \"4885a1b4-8ab5-4d90-934f-c4ca4c3a2347\") " Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.469120 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-utilities" (OuterVolumeSpecName: "utilities") pod "4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" (UID: "4885a1b4-8ab5-4d90-934f-c4ca4c3a2347"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.475251 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-kube-api-access-jd4tn" (OuterVolumeSpecName: "kube-api-access-jd4tn") pod "4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" (UID: "4885a1b4-8ab5-4d90-934f-c4ca4c3a2347"). InnerVolumeSpecName "kube-api-access-jd4tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.571193 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd4tn\" (UniqueName: \"kubernetes.io/projected/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-kube-api-access-jd4tn\") on node \"crc\" DevicePath \"\"" Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.571220 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.571726 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" (UID: "4885a1b4-8ab5-4d90-934f-c4ca4c3a2347"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:16:37 crc kubenswrapper[4931]: I1129 01:16:37.673172 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 01:16:38 crc kubenswrapper[4931]: I1129 01:16:38.400630 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hjzww" Nov 29 01:16:38 crc kubenswrapper[4931]: I1129 01:16:38.434119 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hjzww"] Nov 29 01:16:38 crc kubenswrapper[4931]: I1129 01:16:38.441525 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hjzww"] Nov 29 01:16:39 crc kubenswrapper[4931]: I1129 01:16:39.230914 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" path="/var/lib/kubelet/pods/4885a1b4-8ab5-4d90-934f-c4ca4c3a2347/volumes" Nov 29 01:16:46 crc kubenswrapper[4931]: I1129 01:16:46.213482 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:16:46 crc kubenswrapper[4931]: E1129 01:16:46.214660 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:16:51 crc kubenswrapper[4931]: I1129 01:16:51.027152 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/util/0.log" Nov 29 01:16:51 crc kubenswrapper[4931]: I1129 01:16:51.222717 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/util/0.log" Nov 29 01:16:51 crc kubenswrapper[4931]: I1129 01:16:51.271183 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/pull/0.log" Nov 29 01:16:51 crc kubenswrapper[4931]: I1129 01:16:51.282456 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/pull/0.log" Nov 29 01:16:51 crc kubenswrapper[4931]: I1129 01:16:51.440631 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/util/0.log" Nov 29 01:16:51 crc kubenswrapper[4931]: I1129 01:16:51.445251 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/extract/0.log" Nov 29 01:16:51 crc kubenswrapper[4931]: I1129 01:16:51.528103 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frghcg_990bc8d0-c4bf-4b25-acfb-87dcfde9dc00/pull/0.log" Nov 29 01:16:51 crc kubenswrapper[4931]: I1129 01:16:51.618638 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/util/0.log" Nov 29 01:16:51 crc kubenswrapper[4931]: I1129 01:16:51.787869 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/pull/0.log" Nov 29 01:16:51 crc kubenswrapper[4931]: I1129 01:16:51.800471 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/util/0.log" Nov 29 01:16:51 crc kubenswrapper[4931]: I1129 01:16:51.828853 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/pull/0.log" Nov 29 01:16:52 crc kubenswrapper[4931]: I1129 01:16:52.059002 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/util/0.log" Nov 29 01:16:52 crc kubenswrapper[4931]: I1129 01:16:52.064205 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/extract/0.log" Nov 29 01:16:52 crc kubenswrapper[4931]: I1129 01:16:52.067276 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ddf44_86b012aa-97b4-461d-abf9-19a1a261b806/pull/0.log" Nov 29 01:16:52 crc kubenswrapper[4931]: I1129 01:16:52.249758 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/extract-utilities/0.log" Nov 29 01:16:52 crc kubenswrapper[4931]: I1129 01:16:52.419521 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/extract-utilities/0.log" Nov 29 01:16:52 crc kubenswrapper[4931]: I1129 01:16:52.433689 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/extract-content/0.log" Nov 29 01:16:52 crc kubenswrapper[4931]: I1129 01:16:52.473149 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/extract-content/0.log" Nov 29 01:16:52 crc kubenswrapper[4931]: I1129 01:16:52.605285 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/extract-content/0.log" Nov 29 01:16:52 crc kubenswrapper[4931]: I1129 01:16:52.631892 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/extract-utilities/0.log" Nov 29 01:16:52 crc kubenswrapper[4931]: I1129 01:16:52.839735 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6tsqg_488a91e3-6158-4bae-9210-a3c04ddfc8ac/registry-server/0.log" Nov 29 01:16:52 crc kubenswrapper[4931]: I1129 01:16:52.853605 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gcgfx_531e6518-ff29-45b8-9321-29abd1cf6ec7/extract-utilities/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.042445 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gcgfx_531e6518-ff29-45b8-9321-29abd1cf6ec7/extract-content/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.057226 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gcgfx_531e6518-ff29-45b8-9321-29abd1cf6ec7/extract-content/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.064259 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gcgfx_531e6518-ff29-45b8-9321-29abd1cf6ec7/extract-utilities/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.234438 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gcgfx_531e6518-ff29-45b8-9321-29abd1cf6ec7/extract-utilities/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.276973 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gcgfx_531e6518-ff29-45b8-9321-29abd1cf6ec7/extract-content/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.399951 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gcgfx_531e6518-ff29-45b8-9321-29abd1cf6ec7/registry-server/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.498294 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-m8pmr_db3078a6-11fa-4bda-8b8e-d3ac07d1db11/marketplace-operator/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.551570 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/extract-utilities/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.702460 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/extract-content/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.709107 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/extract-utilities/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.763352 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/extract-content/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.862695 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/extract-utilities/0.log" Nov 29 01:16:53 crc kubenswrapper[4931]: I1129 01:16:53.885630 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/extract-content/0.log" Nov 29 01:16:54 crc kubenswrapper[4931]: I1129 01:16:54.041433 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jwvrh_b3a10cb9-e75c-45e6-a51b-4b6fbeb893c6/registry-server/0.log" Nov 29 01:16:54 crc kubenswrapper[4931]: I1129 01:16:54.098520 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/extract-utilities/0.log" Nov 29 01:16:54 crc kubenswrapper[4931]: I1129 01:16:54.199772 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/extract-content/0.log" Nov 29 01:16:54 crc kubenswrapper[4931]: I1129 01:16:54.214350 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/extract-utilities/0.log" Nov 29 01:16:54 crc kubenswrapper[4931]: I1129 01:16:54.256498 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/extract-content/0.log" Nov 29 01:16:54 crc kubenswrapper[4931]: I1129 01:16:54.392963 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/extract-utilities/0.log" Nov 29 01:16:54 crc kubenswrapper[4931]: I1129 01:16:54.411039 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/extract-content/0.log" Nov 29 01:16:54 crc kubenswrapper[4931]: I1129 01:16:54.953736 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-28xxs_c31367fe-1dab-47be-906b-9324307948d0/registry-server/0.log" Nov 29 01:16:57 crc kubenswrapper[4931]: I1129 01:16:57.220266 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:16:57 crc kubenswrapper[4931]: E1129 01:16:57.225618 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:17:11 crc kubenswrapper[4931]: I1129 01:17:11.217351 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:17:11 crc kubenswrapper[4931]: E1129 01:17:11.218240 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:17:22 crc kubenswrapper[4931]: I1129 01:17:22.212826 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:17:22 crc kubenswrapper[4931]: E1129 01:17:22.213539 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:17:37 crc kubenswrapper[4931]: I1129 01:17:37.227967 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:17:37 crc kubenswrapper[4931]: E1129 01:17:37.229227 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:17:48 crc kubenswrapper[4931]: I1129 01:17:48.212263 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:17:48 crc kubenswrapper[4931]: E1129 01:17:48.213101 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.091062 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qpvsf"] Nov 29 01:17:54 crc kubenswrapper[4931]: E1129 01:17:54.092143 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe6655fe-6ac3-415f-a87d-6a85766fc3fd" containerName="extract-content" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.092155 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe6655fe-6ac3-415f-a87d-6a85766fc3fd" containerName="extract-content" Nov 29 01:17:54 crc kubenswrapper[4931]: E1129 01:17:54.092170 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe6655fe-6ac3-415f-a87d-6a85766fc3fd" containerName="extract-utilities" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.092176 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe6655fe-6ac3-415f-a87d-6a85766fc3fd" containerName="extract-utilities" Nov 29 01:17:54 crc kubenswrapper[4931]: E1129 01:17:54.092193 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe6655fe-6ac3-415f-a87d-6a85766fc3fd" containerName="registry-server" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.092201 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe6655fe-6ac3-415f-a87d-6a85766fc3fd" containerName="registry-server" Nov 29 01:17:54 crc kubenswrapper[4931]: E1129 01:17:54.092213 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" containerName="registry-server" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.092219 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" containerName="registry-server" Nov 29 01:17:54 crc kubenswrapper[4931]: E1129 01:17:54.092229 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" containerName="extract-utilities" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.092235 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" containerName="extract-utilities" Nov 29 01:17:54 crc kubenswrapper[4931]: E1129 01:17:54.092263 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" containerName="extract-content" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.092268 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" containerName="extract-content" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.092423 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="4885a1b4-8ab5-4d90-934f-c4ca4c3a2347" containerName="registry-server" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.092452 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe6655fe-6ac3-415f-a87d-6a85766fc3fd" containerName="registry-server" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.093831 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.124842 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qpvsf"] Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.196322 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84053816-1770-40c6-b468-5212ae499d92-utilities\") pod \"certified-operators-qpvsf\" (UID: \"84053816-1770-40c6-b468-5212ae499d92\") " pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.196403 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84053816-1770-40c6-b468-5212ae499d92-catalog-content\") pod \"certified-operators-qpvsf\" (UID: \"84053816-1770-40c6-b468-5212ae499d92\") " pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.196487 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k5vm\" (UniqueName: \"kubernetes.io/projected/84053816-1770-40c6-b468-5212ae499d92-kube-api-access-8k5vm\") pod \"certified-operators-qpvsf\" (UID: \"84053816-1770-40c6-b468-5212ae499d92\") " pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.297684 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84053816-1770-40c6-b468-5212ae499d92-utilities\") pod \"certified-operators-qpvsf\" (UID: \"84053816-1770-40c6-b468-5212ae499d92\") " pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.298502 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84053816-1770-40c6-b468-5212ae499d92-utilities\") pod \"certified-operators-qpvsf\" (UID: \"84053816-1770-40c6-b468-5212ae499d92\") " pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.300231 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84053816-1770-40c6-b468-5212ae499d92-catalog-content\") pod \"certified-operators-qpvsf\" (UID: \"84053816-1770-40c6-b468-5212ae499d92\") " pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.300364 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84053816-1770-40c6-b468-5212ae499d92-catalog-content\") pod \"certified-operators-qpvsf\" (UID: \"84053816-1770-40c6-b468-5212ae499d92\") " pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.300934 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k5vm\" (UniqueName: \"kubernetes.io/projected/84053816-1770-40c6-b468-5212ae499d92-kube-api-access-8k5vm\") pod \"certified-operators-qpvsf\" (UID: \"84053816-1770-40c6-b468-5212ae499d92\") " pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.324375 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k5vm\" (UniqueName: \"kubernetes.io/projected/84053816-1770-40c6-b468-5212ae499d92-kube-api-access-8k5vm\") pod \"certified-operators-qpvsf\" (UID: \"84053816-1770-40c6-b468-5212ae499d92\") " pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.424498 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:17:54 crc kubenswrapper[4931]: I1129 01:17:54.897326 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qpvsf"] Nov 29 01:17:55 crc kubenswrapper[4931]: I1129 01:17:55.636598 4931 generic.go:334] "Generic (PLEG): container finished" podID="84053816-1770-40c6-b468-5212ae499d92" containerID="aec320bfcc56167dc6b2f6ac2b27a782e4a7353b405c594f15d18a4fecf5ee6b" exitCode=0 Nov 29 01:17:55 crc kubenswrapper[4931]: I1129 01:17:55.636947 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpvsf" event={"ID":"84053816-1770-40c6-b468-5212ae499d92","Type":"ContainerDied","Data":"aec320bfcc56167dc6b2f6ac2b27a782e4a7353b405c594f15d18a4fecf5ee6b"} Nov 29 01:17:55 crc kubenswrapper[4931]: I1129 01:17:55.636978 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpvsf" event={"ID":"84053816-1770-40c6-b468-5212ae499d92","Type":"ContainerStarted","Data":"090d8ede0c5f9eb0764ba2bcbd5a6c1d92d585c1aa89efba4e618475ded95781"} Nov 29 01:17:56 crc kubenswrapper[4931]: I1129 01:17:56.662251 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpvsf" event={"ID":"84053816-1770-40c6-b468-5212ae499d92","Type":"ContainerStarted","Data":"065cbaef0cdf86c93a7d19dc9f14cf7b74d14dbcda1ce1d427ae583db8dfadc0"} Nov 29 01:17:57 crc kubenswrapper[4931]: I1129 01:17:57.685570 4931 generic.go:334] "Generic (PLEG): container finished" podID="84053816-1770-40c6-b468-5212ae499d92" containerID="065cbaef0cdf86c93a7d19dc9f14cf7b74d14dbcda1ce1d427ae583db8dfadc0" exitCode=0 Nov 29 01:17:57 crc kubenswrapper[4931]: I1129 01:17:57.685708 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpvsf" event={"ID":"84053816-1770-40c6-b468-5212ae499d92","Type":"ContainerDied","Data":"065cbaef0cdf86c93a7d19dc9f14cf7b74d14dbcda1ce1d427ae583db8dfadc0"} Nov 29 01:17:58 crc kubenswrapper[4931]: I1129 01:17:58.698182 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpvsf" event={"ID":"84053816-1770-40c6-b468-5212ae499d92","Type":"ContainerStarted","Data":"2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea"} Nov 29 01:17:58 crc kubenswrapper[4931]: I1129 01:17:58.729707 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qpvsf" podStartSLOduration=2.186803606 podStartE2EDuration="4.729688947s" podCreationTimestamp="2025-11-29 01:17:54 +0000 UTC" firstStartedPulling="2025-11-29 01:17:55.640156739 +0000 UTC m=+4688.802049971" lastFinishedPulling="2025-11-29 01:17:58.18304205 +0000 UTC m=+4691.344935312" observedRunningTime="2025-11-29 01:17:58.721419924 +0000 UTC m=+4691.883313206" watchObservedRunningTime="2025-11-29 01:17:58.729688947 +0000 UTC m=+4691.891582189" Nov 29 01:18:01 crc kubenswrapper[4931]: I1129 01:18:01.216068 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:18:01 crc kubenswrapper[4931]: E1129 01:18:01.216985 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:18:04 crc kubenswrapper[4931]: I1129 01:18:04.425588 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:18:04 crc kubenswrapper[4931]: I1129 01:18:04.427076 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:18:04 crc kubenswrapper[4931]: I1129 01:18:04.499923 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:18:04 crc kubenswrapper[4931]: I1129 01:18:04.810444 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:18:04 crc kubenswrapper[4931]: I1129 01:18:04.861906 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qpvsf"] Nov 29 01:18:06 crc kubenswrapper[4931]: I1129 01:18:06.780881 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qpvsf" podUID="84053816-1770-40c6-b468-5212ae499d92" containerName="registry-server" containerID="cri-o://2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea" gracePeriod=2 Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.251956 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.394420 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84053816-1770-40c6-b468-5212ae499d92-catalog-content\") pod \"84053816-1770-40c6-b468-5212ae499d92\" (UID: \"84053816-1770-40c6-b468-5212ae499d92\") " Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.394686 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k5vm\" (UniqueName: \"kubernetes.io/projected/84053816-1770-40c6-b468-5212ae499d92-kube-api-access-8k5vm\") pod \"84053816-1770-40c6-b468-5212ae499d92\" (UID: \"84053816-1770-40c6-b468-5212ae499d92\") " Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.394722 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84053816-1770-40c6-b468-5212ae499d92-utilities\") pod \"84053816-1770-40c6-b468-5212ae499d92\" (UID: \"84053816-1770-40c6-b468-5212ae499d92\") " Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.395551 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84053816-1770-40c6-b468-5212ae499d92-utilities" (OuterVolumeSpecName: "utilities") pod "84053816-1770-40c6-b468-5212ae499d92" (UID: "84053816-1770-40c6-b468-5212ae499d92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.395829 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84053816-1770-40c6-b468-5212ae499d92-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.400767 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84053816-1770-40c6-b468-5212ae499d92-kube-api-access-8k5vm" (OuterVolumeSpecName: "kube-api-access-8k5vm") pod "84053816-1770-40c6-b468-5212ae499d92" (UID: "84053816-1770-40c6-b468-5212ae499d92"). InnerVolumeSpecName "kube-api-access-8k5vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.497764 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k5vm\" (UniqueName: \"kubernetes.io/projected/84053816-1770-40c6-b468-5212ae499d92-kube-api-access-8k5vm\") on node \"crc\" DevicePath \"\"" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.501956 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84053816-1770-40c6-b468-5212ae499d92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84053816-1770-40c6-b468-5212ae499d92" (UID: "84053816-1770-40c6-b468-5212ae499d92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.599589 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84053816-1770-40c6-b468-5212ae499d92-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.794731 4931 generic.go:334] "Generic (PLEG): container finished" podID="84053816-1770-40c6-b468-5212ae499d92" containerID="2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea" exitCode=0 Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.794848 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qpvsf" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.794854 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpvsf" event={"ID":"84053816-1770-40c6-b468-5212ae499d92","Type":"ContainerDied","Data":"2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea"} Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.794981 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qpvsf" event={"ID":"84053816-1770-40c6-b468-5212ae499d92","Type":"ContainerDied","Data":"090d8ede0c5f9eb0764ba2bcbd5a6c1d92d585c1aa89efba4e618475ded95781"} Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.795014 4931 scope.go:117] "RemoveContainer" containerID="2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.826777 4931 scope.go:117] "RemoveContainer" containerID="065cbaef0cdf86c93a7d19dc9f14cf7b74d14dbcda1ce1d427ae583db8dfadc0" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.837793 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qpvsf"] Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.846442 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qpvsf"] Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.853873 4931 scope.go:117] "RemoveContainer" containerID="aec320bfcc56167dc6b2f6ac2b27a782e4a7353b405c594f15d18a4fecf5ee6b" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.895467 4931 scope.go:117] "RemoveContainer" containerID="2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea" Nov 29 01:18:07 crc kubenswrapper[4931]: E1129 01:18:07.896286 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea\": container with ID starting with 2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea not found: ID does not exist" containerID="2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.896329 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea"} err="failed to get container status \"2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea\": rpc error: code = NotFound desc = could not find container \"2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea\": container with ID starting with 2ea078ffbf5ced19c2a618d0edc14c1ca4208829de7c78a6e0b883ae2c79b0ea not found: ID does not exist" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.896355 4931 scope.go:117] "RemoveContainer" containerID="065cbaef0cdf86c93a7d19dc9f14cf7b74d14dbcda1ce1d427ae583db8dfadc0" Nov 29 01:18:07 crc kubenswrapper[4931]: E1129 01:18:07.896867 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"065cbaef0cdf86c93a7d19dc9f14cf7b74d14dbcda1ce1d427ae583db8dfadc0\": container with ID starting with 065cbaef0cdf86c93a7d19dc9f14cf7b74d14dbcda1ce1d427ae583db8dfadc0 not found: ID does not exist" containerID="065cbaef0cdf86c93a7d19dc9f14cf7b74d14dbcda1ce1d427ae583db8dfadc0" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.896896 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"065cbaef0cdf86c93a7d19dc9f14cf7b74d14dbcda1ce1d427ae583db8dfadc0"} err="failed to get container status \"065cbaef0cdf86c93a7d19dc9f14cf7b74d14dbcda1ce1d427ae583db8dfadc0\": rpc error: code = NotFound desc = could not find container \"065cbaef0cdf86c93a7d19dc9f14cf7b74d14dbcda1ce1d427ae583db8dfadc0\": container with ID starting with 065cbaef0cdf86c93a7d19dc9f14cf7b74d14dbcda1ce1d427ae583db8dfadc0 not found: ID does not exist" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.896914 4931 scope.go:117] "RemoveContainer" containerID="aec320bfcc56167dc6b2f6ac2b27a782e4a7353b405c594f15d18a4fecf5ee6b" Nov 29 01:18:07 crc kubenswrapper[4931]: E1129 01:18:07.897335 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aec320bfcc56167dc6b2f6ac2b27a782e4a7353b405c594f15d18a4fecf5ee6b\": container with ID starting with aec320bfcc56167dc6b2f6ac2b27a782e4a7353b405c594f15d18a4fecf5ee6b not found: ID does not exist" containerID="aec320bfcc56167dc6b2f6ac2b27a782e4a7353b405c594f15d18a4fecf5ee6b" Nov 29 01:18:07 crc kubenswrapper[4931]: I1129 01:18:07.897366 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aec320bfcc56167dc6b2f6ac2b27a782e4a7353b405c594f15d18a4fecf5ee6b"} err="failed to get container status \"aec320bfcc56167dc6b2f6ac2b27a782e4a7353b405c594f15d18a4fecf5ee6b\": rpc error: code = NotFound desc = could not find container \"aec320bfcc56167dc6b2f6ac2b27a782e4a7353b405c594f15d18a4fecf5ee6b\": container with ID starting with aec320bfcc56167dc6b2f6ac2b27a782e4a7353b405c594f15d18a4fecf5ee6b not found: ID does not exist" Nov 29 01:18:09 crc kubenswrapper[4931]: I1129 01:18:09.231387 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84053816-1770-40c6-b468-5212ae499d92" path="/var/lib/kubelet/pods/84053816-1770-40c6-b468-5212ae499d92/volumes" Nov 29 01:18:12 crc kubenswrapper[4931]: I1129 01:18:12.214986 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:18:12 crc kubenswrapper[4931]: E1129 01:18:12.215459 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:18:25 crc kubenswrapper[4931]: I1129 01:18:25.216195 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:18:25 crc kubenswrapper[4931]: E1129 01:18:25.217797 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:18:34 crc kubenswrapper[4931]: I1129 01:18:34.148930 4931 generic.go:334] "Generic (PLEG): container finished" podID="143aa3c5-6047-4d24-affe-698ec54aae1b" containerID="f15c4ff22dfd93864b6f6873621985d5fbcdfa3122c761eaef76fc133ba36ae3" exitCode=0 Nov 29 01:18:34 crc kubenswrapper[4931]: I1129 01:18:34.149056 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zn82b/must-gather-gtd7t" event={"ID":"143aa3c5-6047-4d24-affe-698ec54aae1b","Type":"ContainerDied","Data":"f15c4ff22dfd93864b6f6873621985d5fbcdfa3122c761eaef76fc133ba36ae3"} Nov 29 01:18:34 crc kubenswrapper[4931]: I1129 01:18:34.150309 4931 scope.go:117] "RemoveContainer" containerID="f15c4ff22dfd93864b6f6873621985d5fbcdfa3122c761eaef76fc133ba36ae3" Nov 29 01:18:34 crc kubenswrapper[4931]: I1129 01:18:34.890446 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zn82b_must-gather-gtd7t_143aa3c5-6047-4d24-affe-698ec54aae1b/gather/0.log" Nov 29 01:18:38 crc kubenswrapper[4931]: I1129 01:18:38.212965 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:18:38 crc kubenswrapper[4931]: E1129 01:18:38.215044 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:18:44 crc kubenswrapper[4931]: I1129 01:18:44.927420 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zn82b/must-gather-gtd7t"] Nov 29 01:18:44 crc kubenswrapper[4931]: I1129 01:18:44.928270 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-zn82b/must-gather-gtd7t" podUID="143aa3c5-6047-4d24-affe-698ec54aae1b" containerName="copy" containerID="cri-o://7d2931064538572450525d2e1fe19643cdb225a7ea2579dcf8a0009599f85af8" gracePeriod=2 Nov 29 01:18:44 crc kubenswrapper[4931]: I1129 01:18:44.936050 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zn82b/must-gather-gtd7t"] Nov 29 01:18:45 crc kubenswrapper[4931]: I1129 01:18:45.300054 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zn82b_must-gather-gtd7t_143aa3c5-6047-4d24-affe-698ec54aae1b/copy/0.log" Nov 29 01:18:45 crc kubenswrapper[4931]: I1129 01:18:45.301435 4931 generic.go:334] "Generic (PLEG): container finished" podID="143aa3c5-6047-4d24-affe-698ec54aae1b" containerID="7d2931064538572450525d2e1fe19643cdb225a7ea2579dcf8a0009599f85af8" exitCode=143 Nov 29 01:18:45 crc kubenswrapper[4931]: I1129 01:18:45.301658 4931 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e7c7462e436cbe72c15a1a51e6f22a236f8a8fda49988622ad7a7b0feaf5294" Nov 29 01:18:45 crc kubenswrapper[4931]: I1129 01:18:45.346390 4931 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zn82b_must-gather-gtd7t_143aa3c5-6047-4d24-affe-698ec54aae1b/copy/0.log" Nov 29 01:18:45 crc kubenswrapper[4931]: I1129 01:18:45.347157 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/must-gather-gtd7t" Nov 29 01:18:45 crc kubenswrapper[4931]: I1129 01:18:45.525321 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pn828\" (UniqueName: \"kubernetes.io/projected/143aa3c5-6047-4d24-affe-698ec54aae1b-kube-api-access-pn828\") pod \"143aa3c5-6047-4d24-affe-698ec54aae1b\" (UID: \"143aa3c5-6047-4d24-affe-698ec54aae1b\") " Nov 29 01:18:45 crc kubenswrapper[4931]: I1129 01:18:45.526247 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/143aa3c5-6047-4d24-affe-698ec54aae1b-must-gather-output\") pod \"143aa3c5-6047-4d24-affe-698ec54aae1b\" (UID: \"143aa3c5-6047-4d24-affe-698ec54aae1b\") " Nov 29 01:18:45 crc kubenswrapper[4931]: I1129 01:18:45.531288 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/143aa3c5-6047-4d24-affe-698ec54aae1b-kube-api-access-pn828" (OuterVolumeSpecName: "kube-api-access-pn828") pod "143aa3c5-6047-4d24-affe-698ec54aae1b" (UID: "143aa3c5-6047-4d24-affe-698ec54aae1b"). InnerVolumeSpecName "kube-api-access-pn828". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:18:45 crc kubenswrapper[4931]: I1129 01:18:45.628546 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pn828\" (UniqueName: \"kubernetes.io/projected/143aa3c5-6047-4d24-affe-698ec54aae1b-kube-api-access-pn828\") on node \"crc\" DevicePath \"\"" Nov 29 01:18:45 crc kubenswrapper[4931]: I1129 01:18:45.685258 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/143aa3c5-6047-4d24-affe-698ec54aae1b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "143aa3c5-6047-4d24-affe-698ec54aae1b" (UID: "143aa3c5-6047-4d24-affe-698ec54aae1b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:18:45 crc kubenswrapper[4931]: I1129 01:18:45.730084 4931 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/143aa3c5-6047-4d24-affe-698ec54aae1b-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 29 01:18:46 crc kubenswrapper[4931]: I1129 01:18:46.308597 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zn82b/must-gather-gtd7t" Nov 29 01:18:47 crc kubenswrapper[4931]: I1129 01:18:47.240615 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="143aa3c5-6047-4d24-affe-698ec54aae1b" path="/var/lib/kubelet/pods/143aa3c5-6047-4d24-affe-698ec54aae1b/volumes" Nov 29 01:18:50 crc kubenswrapper[4931]: I1129 01:18:50.212225 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:18:50 crc kubenswrapper[4931]: E1129 01:18:50.212927 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:19:02 crc kubenswrapper[4931]: I1129 01:19:02.212995 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:19:02 crc kubenswrapper[4931]: E1129 01:19:02.214169 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:19:14 crc kubenswrapper[4931]: I1129 01:19:14.212785 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:19:14 crc kubenswrapper[4931]: E1129 01:19:14.213942 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:19:28 crc kubenswrapper[4931]: I1129 01:19:28.212711 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:19:28 crc kubenswrapper[4931]: E1129 01:19:28.213479 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:19:40 crc kubenswrapper[4931]: I1129 01:19:40.212387 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:19:40 crc kubenswrapper[4931]: E1129 01:19:40.213355 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:19:54 crc kubenswrapper[4931]: I1129 01:19:54.212273 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:19:54 crc kubenswrapper[4931]: E1129 01:19:54.213354 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:20:07 crc kubenswrapper[4931]: I1129 01:20:07.227417 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:20:07 crc kubenswrapper[4931]: E1129 01:20:07.228746 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:20:18 crc kubenswrapper[4931]: I1129 01:20:18.247363 4931 scope.go:117] "RemoveContainer" containerID="f15c4ff22dfd93864b6f6873621985d5fbcdfa3122c761eaef76fc133ba36ae3" Nov 29 01:20:18 crc kubenswrapper[4931]: I1129 01:20:18.360296 4931 scope.go:117] "RemoveContainer" containerID="6930cc45ec83f44b152fedfbab1ba38e3eb93d4d9d938bab59f66e48da500a23" Nov 29 01:20:18 crc kubenswrapper[4931]: I1129 01:20:18.394508 4931 scope.go:117] "RemoveContainer" containerID="7d2931064538572450525d2e1fe19643cdb225a7ea2579dcf8a0009599f85af8" Nov 29 01:20:20 crc kubenswrapper[4931]: I1129 01:20:20.212150 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:20:20 crc kubenswrapper[4931]: E1129 01:20:20.212937 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:20:35 crc kubenswrapper[4931]: I1129 01:20:35.215187 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:20:35 crc kubenswrapper[4931]: E1129 01:20:35.216473 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:20:47 crc kubenswrapper[4931]: I1129 01:20:47.238914 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:20:47 crc kubenswrapper[4931]: E1129 01:20:47.240314 4931 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6xxhj_openshift-machine-config-operator(0b227491-f635-4817-96ed-90b96ab6aab3)\"" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" podUID="0b227491-f635-4817-96ed-90b96ab6aab3" Nov 29 01:20:59 crc kubenswrapper[4931]: I1129 01:20:59.213485 4931 scope.go:117] "RemoveContainer" containerID="11ddfb649a0d9eae3ba6ee0094520f66f470bfe1b6234b06f5e4d926f5a0d35a" Nov 29 01:20:59 crc kubenswrapper[4931]: I1129 01:20:59.900494 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6xxhj" event={"ID":"0b227491-f635-4817-96ed-90b96ab6aab3","Type":"ContainerStarted","Data":"3a513b0e810237c81df68a6b09ec09b392722933615474afcfc3043568d55018"} Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.699292 4931 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jtqw8"] Nov 29 01:22:04 crc kubenswrapper[4931]: E1129 01:22:04.700257 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84053816-1770-40c6-b468-5212ae499d92" containerName="extract-utilities" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.700274 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="84053816-1770-40c6-b468-5212ae499d92" containerName="extract-utilities" Nov 29 01:22:04 crc kubenswrapper[4931]: E1129 01:22:04.700306 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143aa3c5-6047-4d24-affe-698ec54aae1b" containerName="copy" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.700314 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="143aa3c5-6047-4d24-affe-698ec54aae1b" containerName="copy" Nov 29 01:22:04 crc kubenswrapper[4931]: E1129 01:22:04.700328 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84053816-1770-40c6-b468-5212ae499d92" containerName="extract-content" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.700337 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="84053816-1770-40c6-b468-5212ae499d92" containerName="extract-content" Nov 29 01:22:04 crc kubenswrapper[4931]: E1129 01:22:04.700353 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143aa3c5-6047-4d24-affe-698ec54aae1b" containerName="gather" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.700360 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="143aa3c5-6047-4d24-affe-698ec54aae1b" containerName="gather" Nov 29 01:22:04 crc kubenswrapper[4931]: E1129 01:22:04.700385 4931 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84053816-1770-40c6-b468-5212ae499d92" containerName="registry-server" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.700393 4931 state_mem.go:107] "Deleted CPUSet assignment" podUID="84053816-1770-40c6-b468-5212ae499d92" containerName="registry-server" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.700609 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="84053816-1770-40c6-b468-5212ae499d92" containerName="registry-server" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.700642 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="143aa3c5-6047-4d24-affe-698ec54aae1b" containerName="gather" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.700656 4931 memory_manager.go:354] "RemoveStaleState removing state" podUID="143aa3c5-6047-4d24-affe-698ec54aae1b" containerName="copy" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.702307 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.712229 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jtqw8"] Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.751061 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4jt7\" (UniqueName: \"kubernetes.io/projected/61c4839d-3816-479c-bce7-840bbd003e73-kube-api-access-m4jt7\") pod \"community-operators-jtqw8\" (UID: \"61c4839d-3816-479c-bce7-840bbd003e73\") " pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.751149 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61c4839d-3816-479c-bce7-840bbd003e73-utilities\") pod \"community-operators-jtqw8\" (UID: \"61c4839d-3816-479c-bce7-840bbd003e73\") " pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.751210 4931 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61c4839d-3816-479c-bce7-840bbd003e73-catalog-content\") pod \"community-operators-jtqw8\" (UID: \"61c4839d-3816-479c-bce7-840bbd003e73\") " pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.852373 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61c4839d-3816-479c-bce7-840bbd003e73-catalog-content\") pod \"community-operators-jtqw8\" (UID: \"61c4839d-3816-479c-bce7-840bbd003e73\") " pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.852667 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4jt7\" (UniqueName: \"kubernetes.io/projected/61c4839d-3816-479c-bce7-840bbd003e73-kube-api-access-m4jt7\") pod \"community-operators-jtqw8\" (UID: \"61c4839d-3816-479c-bce7-840bbd003e73\") " pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.852718 4931 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61c4839d-3816-479c-bce7-840bbd003e73-utilities\") pod \"community-operators-jtqw8\" (UID: \"61c4839d-3816-479c-bce7-840bbd003e73\") " pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.852876 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61c4839d-3816-479c-bce7-840bbd003e73-catalog-content\") pod \"community-operators-jtqw8\" (UID: \"61c4839d-3816-479c-bce7-840bbd003e73\") " pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.853031 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61c4839d-3816-479c-bce7-840bbd003e73-utilities\") pod \"community-operators-jtqw8\" (UID: \"61c4839d-3816-479c-bce7-840bbd003e73\") " pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:04 crc kubenswrapper[4931]: I1129 01:22:04.881852 4931 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4jt7\" (UniqueName: \"kubernetes.io/projected/61c4839d-3816-479c-bce7-840bbd003e73-kube-api-access-m4jt7\") pod \"community-operators-jtqw8\" (UID: \"61c4839d-3816-479c-bce7-840bbd003e73\") " pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:05 crc kubenswrapper[4931]: I1129 01:22:05.030491 4931 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:05 crc kubenswrapper[4931]: I1129 01:22:05.630339 4931 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jtqw8"] Nov 29 01:22:05 crc kubenswrapper[4931]: I1129 01:22:05.748525 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtqw8" event={"ID":"61c4839d-3816-479c-bce7-840bbd003e73","Type":"ContainerStarted","Data":"5925bc215909ddeaa2fe79fdd8659d2bed63968d999d0ae87ade158cc3cca5c0"} Nov 29 01:22:06 crc kubenswrapper[4931]: I1129 01:22:06.757770 4931 generic.go:334] "Generic (PLEG): container finished" podID="61c4839d-3816-479c-bce7-840bbd003e73" containerID="665a7c15d77953ed225b0f3402bb75b8463fbb28dafc43aa058c34cdfaacef63" exitCode=0 Nov 29 01:22:06 crc kubenswrapper[4931]: I1129 01:22:06.757858 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtqw8" event={"ID":"61c4839d-3816-479c-bce7-840bbd003e73","Type":"ContainerDied","Data":"665a7c15d77953ed225b0f3402bb75b8463fbb28dafc43aa058c34cdfaacef63"} Nov 29 01:22:06 crc kubenswrapper[4931]: I1129 01:22:06.760150 4931 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 29 01:22:07 crc kubenswrapper[4931]: I1129 01:22:07.769323 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtqw8" event={"ID":"61c4839d-3816-479c-bce7-840bbd003e73","Type":"ContainerStarted","Data":"00e089f0c2faf5100666af3646b5145a94689566075aa1450ddbb56d2a8a8d41"} Nov 29 01:22:08 crc kubenswrapper[4931]: I1129 01:22:08.783843 4931 generic.go:334] "Generic (PLEG): container finished" podID="61c4839d-3816-479c-bce7-840bbd003e73" containerID="00e089f0c2faf5100666af3646b5145a94689566075aa1450ddbb56d2a8a8d41" exitCode=0 Nov 29 01:22:08 crc kubenswrapper[4931]: I1129 01:22:08.783929 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtqw8" event={"ID":"61c4839d-3816-479c-bce7-840bbd003e73","Type":"ContainerDied","Data":"00e089f0c2faf5100666af3646b5145a94689566075aa1450ddbb56d2a8a8d41"} Nov 29 01:22:09 crc kubenswrapper[4931]: I1129 01:22:09.796336 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtqw8" event={"ID":"61c4839d-3816-479c-bce7-840bbd003e73","Type":"ContainerStarted","Data":"697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c"} Nov 29 01:22:09 crc kubenswrapper[4931]: I1129 01:22:09.820217 4931 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jtqw8" podStartSLOduration=3.327499951 podStartE2EDuration="5.820196629s" podCreationTimestamp="2025-11-29 01:22:04 +0000 UTC" firstStartedPulling="2025-11-29 01:22:06.7599394 +0000 UTC m=+4939.921832632" lastFinishedPulling="2025-11-29 01:22:09.252636068 +0000 UTC m=+4942.414529310" observedRunningTime="2025-11-29 01:22:09.81774955 +0000 UTC m=+4942.979642812" watchObservedRunningTime="2025-11-29 01:22:09.820196629 +0000 UTC m=+4942.982089871" Nov 29 01:22:15 crc kubenswrapper[4931]: I1129 01:22:15.030720 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:15 crc kubenswrapper[4931]: I1129 01:22:15.031205 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:15 crc kubenswrapper[4931]: I1129 01:22:15.103765 4931 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:15 crc kubenswrapper[4931]: I1129 01:22:15.943445 4931 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:15 crc kubenswrapper[4931]: I1129 01:22:15.994919 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jtqw8"] Nov 29 01:22:17 crc kubenswrapper[4931]: I1129 01:22:17.875517 4931 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jtqw8" podUID="61c4839d-3816-479c-bce7-840bbd003e73" containerName="registry-server" containerID="cri-o://697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c" gracePeriod=2 Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.446210 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.547022 4931 scope.go:117] "RemoveContainer" containerID="bef9ee83f4f2be5571f73f631bab1f388fe681a54d0a098984c6688b41dd2a25" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.553022 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61c4839d-3816-479c-bce7-840bbd003e73-utilities\") pod \"61c4839d-3816-479c-bce7-840bbd003e73\" (UID: \"61c4839d-3816-479c-bce7-840bbd003e73\") " Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.553080 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61c4839d-3816-479c-bce7-840bbd003e73-catalog-content\") pod \"61c4839d-3816-479c-bce7-840bbd003e73\" (UID: \"61c4839d-3816-479c-bce7-840bbd003e73\") " Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.553245 4931 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4jt7\" (UniqueName: \"kubernetes.io/projected/61c4839d-3816-479c-bce7-840bbd003e73-kube-api-access-m4jt7\") pod \"61c4839d-3816-479c-bce7-840bbd003e73\" (UID: \"61c4839d-3816-479c-bce7-840bbd003e73\") " Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.554721 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61c4839d-3816-479c-bce7-840bbd003e73-utilities" (OuterVolumeSpecName: "utilities") pod "61c4839d-3816-479c-bce7-840bbd003e73" (UID: "61c4839d-3816-479c-bce7-840bbd003e73"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.564549 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61c4839d-3816-479c-bce7-840bbd003e73-kube-api-access-m4jt7" (OuterVolumeSpecName: "kube-api-access-m4jt7") pod "61c4839d-3816-479c-bce7-840bbd003e73" (UID: "61c4839d-3816-479c-bce7-840bbd003e73"). InnerVolumeSpecName "kube-api-access-m4jt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.587754 4931 scope.go:117] "RemoveContainer" containerID="71293f474e38d99d0fa4d588bd729a6793948fd542da05f4267e14ba869b6c8c" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.647363 4931 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61c4839d-3816-479c-bce7-840bbd003e73-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61c4839d-3816-479c-bce7-840bbd003e73" (UID: "61c4839d-3816-479c-bce7-840bbd003e73"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.654206 4931 scope.go:117] "RemoveContainer" containerID="c764a3dd78de84f71cf1a2682d0c9b42c7a8910b7e8a98eb0184cb523d9189c9" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.655159 4931 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61c4839d-3816-479c-bce7-840bbd003e73-utilities\") on node \"crc\" DevicePath \"\"" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.655191 4931 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61c4839d-3816-479c-bce7-840bbd003e73-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.655202 4931 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4jt7\" (UniqueName: \"kubernetes.io/projected/61c4839d-3816-479c-bce7-840bbd003e73-kube-api-access-m4jt7\") on node \"crc\" DevicePath \"\"" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.891705 4931 generic.go:334] "Generic (PLEG): container finished" podID="61c4839d-3816-479c-bce7-840bbd003e73" containerID="697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c" exitCode=0 Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.891785 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtqw8" event={"ID":"61c4839d-3816-479c-bce7-840bbd003e73","Type":"ContainerDied","Data":"697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c"} Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.891882 4931 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtqw8" event={"ID":"61c4839d-3816-479c-bce7-840bbd003e73","Type":"ContainerDied","Data":"5925bc215909ddeaa2fe79fdd8659d2bed63968d999d0ae87ade158cc3cca5c0"} Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.891935 4931 scope.go:117] "RemoveContainer" containerID="697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.891941 4931 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jtqw8" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.917744 4931 scope.go:117] "RemoveContainer" containerID="00e089f0c2faf5100666af3646b5145a94689566075aa1450ddbb56d2a8a8d41" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.959531 4931 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jtqw8"] Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.964975 4931 scope.go:117] "RemoveContainer" containerID="665a7c15d77953ed225b0f3402bb75b8463fbb28dafc43aa058c34cdfaacef63" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.973282 4931 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jtqw8"] Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.993038 4931 scope.go:117] "RemoveContainer" containerID="697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c" Nov 29 01:22:18 crc kubenswrapper[4931]: E1129 01:22:18.996455 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c\": container with ID starting with 697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c not found: ID does not exist" containerID="697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.996500 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c"} err="failed to get container status \"697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c\": rpc error: code = NotFound desc = could not find container \"697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c\": container with ID starting with 697658469c14e2faea6fed1a0090ad7454c1e8c32a650044d25036fc33bbc75c not found: ID does not exist" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.996525 4931 scope.go:117] "RemoveContainer" containerID="00e089f0c2faf5100666af3646b5145a94689566075aa1450ddbb56d2a8a8d41" Nov 29 01:22:18 crc kubenswrapper[4931]: E1129 01:22:18.997200 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00e089f0c2faf5100666af3646b5145a94689566075aa1450ddbb56d2a8a8d41\": container with ID starting with 00e089f0c2faf5100666af3646b5145a94689566075aa1450ddbb56d2a8a8d41 not found: ID does not exist" containerID="00e089f0c2faf5100666af3646b5145a94689566075aa1450ddbb56d2a8a8d41" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.997231 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00e089f0c2faf5100666af3646b5145a94689566075aa1450ddbb56d2a8a8d41"} err="failed to get container status \"00e089f0c2faf5100666af3646b5145a94689566075aa1450ddbb56d2a8a8d41\": rpc error: code = NotFound desc = could not find container \"00e089f0c2faf5100666af3646b5145a94689566075aa1450ddbb56d2a8a8d41\": container with ID starting with 00e089f0c2faf5100666af3646b5145a94689566075aa1450ddbb56d2a8a8d41 not found: ID does not exist" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.997258 4931 scope.go:117] "RemoveContainer" containerID="665a7c15d77953ed225b0f3402bb75b8463fbb28dafc43aa058c34cdfaacef63" Nov 29 01:22:18 crc kubenswrapper[4931]: E1129 01:22:18.997525 4931 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"665a7c15d77953ed225b0f3402bb75b8463fbb28dafc43aa058c34cdfaacef63\": container with ID starting with 665a7c15d77953ed225b0f3402bb75b8463fbb28dafc43aa058c34cdfaacef63 not found: ID does not exist" containerID="665a7c15d77953ed225b0f3402bb75b8463fbb28dafc43aa058c34cdfaacef63" Nov 29 01:22:18 crc kubenswrapper[4931]: I1129 01:22:18.997551 4931 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"665a7c15d77953ed225b0f3402bb75b8463fbb28dafc43aa058c34cdfaacef63"} err="failed to get container status \"665a7c15d77953ed225b0f3402bb75b8463fbb28dafc43aa058c34cdfaacef63\": rpc error: code = NotFound desc = could not find container \"665a7c15d77953ed225b0f3402bb75b8463fbb28dafc43aa058c34cdfaacef63\": container with ID starting with 665a7c15d77953ed225b0f3402bb75b8463fbb28dafc43aa058c34cdfaacef63 not found: ID does not exist" Nov 29 01:22:19 crc kubenswrapper[4931]: I1129 01:22:19.226940 4931 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61c4839d-3816-479c-bce7-840bbd003e73" path="/var/lib/kubelet/pods/61c4839d-3816-479c-bce7-840bbd003e73/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515112445337024452 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015112445340017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015112433242016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015112433242015452 5ustar corecore